sha
null
last_modified
null
library_name
stringclasses
154 values
text
stringlengths
1
900k
metadata
stringlengths
2
348k
pipeline_tag
stringclasses
45 values
id
stringlengths
5
122
tags
sequencelengths
1
1.84k
created_at
stringlengths
25
25
arxiv
sequencelengths
0
201
languages
sequencelengths
0
1.83k
tags_str
stringlengths
17
9.34k
text_str
stringlengths
0
389k
text_lists
sequencelengths
0
722
processed_texts
sequencelengths
1
723
tokens_length
sequencelengths
1
723
input_texts
sequencelengths
1
61
embeddings
sequencelengths
768
768
null
null
null
# w2v with news
{}
null
Aviora/news2vec
[ "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #region-us
# w2v with news
[ "# w2v with news" ]
[ "TAGS\n#region-us \n", "# w2v with news" ]
[ 6, 6 ]
[ "passage: TAGS\n#region-us \n# w2v with news" ]
[ 0.03270259499549866, -0.029801558703184128, -0.009985742159187794, -0.09686803817749023, 0.0035065042320638895, 0.07539942860603333, 0.028866833075881004, 0.06558684259653091, 0.15416160225868225, -0.03910063952207565, 0.16416548192501068, -0.05236139893531799, -0.05017606541514397, 0.024272512644529343, 0.007409993093460798, -0.1307477504014969, 0.11534982919692993, -0.07066880911588669, 0.0016109326388686895, 0.06093878298997879, -0.042304087430238724, -0.061699360609054565, -0.009548652917146683, -0.045090895146131516, -0.1096712201833725, 0.09584394842386246, 0.03854630887508392, 0.022607844322919846, 0.10162214934825897, -0.025431368499994278, 0.09155970066785812, 0.034583572298288345, -0.010424106381833553, -0.19961373507976532, 0.02895129844546318, -0.03919507935643196, -0.1007479578256607, 0.042533595114946365, 0.0843077003955841, -0.07529239356517792, -0.1023605614900589, -0.0853671208024025, -0.0848083645105362, 0.06278801709413528, -0.24963778257369995, -0.17205539345741272, -0.005593018140643835, 0.028884639963507652, 0.10296550393104553, 0.0573284812271595, -0.004050846677273512, 0.19999922811985016, -0.18370682001113892, 0.0020747785456478596, 0.15976709127426147, -0.2630353271961212, 0.10363546013832092, 0.06125284731388092, -0.06329837441444397, 0.048543781042099, 0.00983964093029499, 0.1306189000606537, 0.05494399741292, -0.03157433494925499, -0.2416858971118927, -0.02651265822350979, -0.15661421418190002, 0.09091790020465851, -0.060217346996068954, -0.11894815415143967, 0.18602745234966278, 0.05275500565767288, 0.054321322590112686, 0.16193635761737823, -0.07096967846155167, -0.0686551034450531, 0.04809783399105072, 0.03595324978232384, -0.049305204302072525, 0.14060817658901215, 0.13267302513122559, -0.024265943095088005, -0.15921863913536072, 0.012784978374838829, -0.21257351338863373, 0.3187193274497986, -0.031240936368703842, 0.19824838638305664, -0.2570606768131256, 0.054038263857364655, -0.10532760620117188, -0.06923321634531021, 0.04477907717227936, -0.11093372106552124, 0.0310373455286026, 0.04210768640041351, -0.04535531997680664, 0.039309874176979065, 0.014958337880671024, -0.13586632907390594, -0.15064822137355804, 0.0034389030188322067, -0.04079966992139816, 0.19836631417274475, 0.1370328813791275, -0.12588870525360107, -0.0013671095948666334, 0.041693203151226044, -0.039757296442985535, -0.14941798150539398, 0.04270635172724724, -0.033778414130210876, -0.0765921100974083, -0.009441492147743702, -0.18603122234344482, 0.10829851031303406, 0.03721761330962181, -0.06812174618244171, -0.1596231907606125, 0.0871218740940094, 0.056606244295835495, -0.048414625227451324, -0.08233992755413055, 0.021893437951803207, -0.0026117381639778614, 0.14001412689685822, -0.06901408731937408, 0.052161768078804016, 0.0890587717294693, 0.11056900024414062, -0.13564841449260712, -0.04437738284468651, -0.04465864971280098, 0.09041709452867508, 0.0867047980427742, 0.07576655596494675, 0.021467428654432297, -0.060627106577157974, -0.10191644728183746, -0.002099039498716593, -0.1086522787809372, 0.03069615550339222, 0.09839648753404617, -0.01372779905796051, 0.06021658703684807, -0.0059724957682192326, -0.01729193516075611, 0.018134452402591705, -0.09940563887357712, 0.08850111067295074, -0.011965835466980934, 0.05064138025045395, -0.181556835770607, -0.040815308690071106, -0.18570387363433838, 0.11364641040563583, 0.16756480932235718, 0.04829777404665947, -0.05539941415190697, 0.15422379970550537, 0.04825746640563011, 0.07351385056972504, -0.2381320744752884, 0.01909908466041088, -0.06757084280252457, 0.2854062616825104, -0.18002139031887054, -0.10089187324047089, 0.2640584409236908, -0.06278223544359207, 0.03791835904121399, -0.06689700484275818, 0.0010870948899537325, 0.025942210108041763, 0.14117999374866486, 0.25500890612602234, -0.09938384592533112, -0.022871503606438637, 0.0005245289066806436, 0.11009504646062851, -0.1692880094051361, -0.006494434084743261, 0.06720932573080063, 0.014415369369089603, -0.19554269313812256, 0.017006205394864082, 0.06520406901836395, 0.047539666295051575, -0.08123163878917694, -0.05845923721790314, 0.07166268676519394, 0.002615709789097309, 0.18098267912864685, -0.003333097090944648, 0.043536633253097534, -0.07249698787927628, 0.006248024292290211, 0.0516088642179966, 0.10996657609939575, 0.1428365260362625, 0.0498143807053566, -0.08798689395189285, 0.03193838894367218, -0.025089871138334274, 0.07387513667345047, -0.06013687327504158, -0.058094240725040436, -0.043281275779008865, 0.13994091749191284, 0.06238485872745514, 0.23290053009986877, 0.07608118653297424, -0.12975242733955383, -0.008409393951296806, -0.03236667439341545, -0.03761685639619827, 0.05119486153125763, 0.011496756225824356, -0.11380332708358765, 0.08926936239004135, -0.09348910301923752, -0.06653743982315063, -0.14088910818099976, -0.04305162653326988, 0.19112859666347504, 0.05802055820822716, 0.047181107103824615, 0.002772500505670905, 0.03612599894404411, -0.021871842443943024, 0.09521646052598953, 0.027175726369023323, 0.07059358805418015, -0.05809597671031952, -0.06868988275527954, 0.18115292489528656, -0.05620342865586281, 0.17034998536109924, 0.10551628470420837, -0.2405051738023758, 0.0654337927699089, 0.041484445333480835, -0.013302070088684559, 0.005605232436209917, -0.02690241113305092, -0.08641377836465836, -0.024210767820477486, -0.07530990242958069, -0.053554635494947433, 0.03319856524467468, -0.018716245889663696, -0.05020640790462494, -0.018276534974575043, -0.08301036804914474, 0.07635768502950668, 0.027074575424194336, -0.15989968180656433, 0.06655314564704895, 0.4196056127548218, 0.1605922430753708, 0.2560180723667145, -0.04322348162531853, -0.013211260549724102, 0.042461369186639786, -0.11732766777276993, -0.10462486743927002, 0.07275810837745667, -0.07505791634321213, 0.07388262450695038, 0.025258613750338554, 0.006955668330192566, 0.10082494467496872, -0.12232845276594162, -0.13614556193351746, -0.02039952203631401, -0.006531078368425369, -0.2248077094554901, -0.04180743545293808, -0.07914792001247406, 0.08001892268657684, 0.054675571620464325, 0.07827149331569672, 0.09541171044111252, -0.04920712858438492, -0.09781820327043533, 0.015951503068208694, -0.11593735963106155, -0.2272207885980606, -0.12171846628189087, 0.0071153840981423855, 0.08749279379844666, 0.04392256960272789, -0.039645254611968994, -0.1614760011434555, -0.004411192610859871, 0.07041484862565994, 0.09502055495977402, -0.09006210416555405, 0.051029108464717865, 0.19936026632785797, 0.1388636827468872, -0.027089949697256088, -0.03673649951815605, -0.008288168348371983, -0.12473863363265991, -0.0511975958943367, 0.029271718114614487, -0.10141909122467041, 0.0801122635602951, 0.14936664700508118, 0.11202864348888397, 0.05552642047405243, 0.006895365659147501, 0.09297271817922592, -0.10654918849468231, -0.07805568724870682, 0.07954870164394379, -0.15457376837730408, 0.03598189726471901, 0.1667109578847885, 0.09747038036584854, -0.08942226320505142, -0.02110222540795803, -0.04599966108798981, -0.04143228381872177, -0.2777545750141144, -0.0992046371102333, -0.032495126128196716, 0.10867973417043686, -0.11933165043592453, 0.03842305764555931, 0.01486167125403881, -0.031591035425662994, -0.00487865973263979, -0.098752960562706, -0.04282223433256149, 0.005694931838661432, 0.15215520560741425, -0.03396311402320862, -0.04912392795085907, -0.1075807511806488, -0.03428942337632179, 0.11834754794836044, 0.060218628495931625, -0.07633261382579803, 0.17744797468185425, 0.044108301401138306, 0.02163197658956051, 0.020413022488355637, 0.19751043617725372, 0.07014136016368866, 0.09243572503328323, -0.07968450337648392, -0.00350164994597435, -0.0060820309445261955, -0.030060095712542534, 0.04506746307015419, 0.054857924580574036, -0.19319289922714233, 0.06560900807380676, -0.18091844022274017, 0.06848953664302826, -0.028172407299280167, 0.15504337847232819, -0.005851328372955322, 0.07751758396625519, 0.14210930466651917, 0.011896041221916676, -0.04426254332065582, 0.10278382152318954, 0.06867675483226776, -0.06274425238370895, 0.09768988937139511, 0.16385281085968018, 0.09419842064380646, -0.05193585157394409, 0.07250632345676422, -0.14751140773296356, -0.28667739033699036, -0.005309464875608683, 0.15674948692321777, -0.10587918758392334, 0.2735496759414673, 0.03270446136593819, -0.12761329114437103, -0.02285267971456051, -0.16956879198551178, -0.07693497836589813, 0.12962080538272858, 0.08341704308986664, 0.05955738574266434, 0.004042687360197306, -0.02830643206834793, -0.12415552884340286, 0.02647463046014309, 0.09310419112443924, -0.06568224728107452, -0.07632509618997574, 0.03088480979204178, 0.000039324164390563965, -0.01797754131257534, -0.042739834636449814, -0.036411941051483154, 0.04213477671146393, 0.03082464262843132, 0.01637187972664833, 0.0768560841679573, -0.03767096996307373, -0.06061648204922676, -0.012672189623117447, -0.009690731763839722, -0.09582570195198059, 0.033984191715717316, -0.10493738204240799, 0.0410466194152832, -0.013747175224125385, -0.05614287033677101, -0.09760326147079468, 0.0526372566819191, 0.025100450962781906, -0.08320163935422897, -0.1896093338727951, 0.1465754210948944, -0.06378203630447388, -0.10269854217767715, -0.07140778005123138, 0.22936135530471802, 0.0640977993607521, 0.07103350013494492, -0.06672472506761551, -0.005853875074535608, 0.07134580612182617, -0.08861514180898666, 0.19416089355945587, -0.01215298380702734, -0.058417998254299164, 0.09426393359899521, 0.1157398670911789, -0.11423628032207489, -0.00473414920270443, -0.032121289521455765, 0.18881501257419586, 0.40714260935783386, -0.057384926825761795, 0.23324747383594513, 0.19808189570903778, 0.008975183591246605, -0.3198009133338928, -0.07864539325237274, -0.13186876475811005, -0.04233001172542572, 0.1490900218486786, -0.12008312344551086, 0.09291037917137146, 0.10859408974647522, -0.08788181096315384, 0.3049028813838959, -0.3155151307582855, -0.017929386347532272, -0.0017042001709342003, -0.20213375985622406, 0.6906268000602722, -0.032215941697359085, -0.11852703988552094, 0.019951431080698967, -0.1645057201385498, 0.203105628490448, -0.08416932821273804, -0.0017794298473745584, 0.008947933092713356, 0.058374401181936264, 0.03942878916859627, 0.0018215819727629423, 0.16660962998867035, 0.009943793527781963, 0.007026787381619215, -0.08478652685880661, -0.11153558641672134, 0.182327002286911, 0.0168877225369215, -0.16555260121822357, 0.07552304118871689, -0.010012039914727211, -0.19697536528110504, 0.00009728828445076942, -0.09168512374162674, 0.010677826590836048, 0.06818314641714096, 0.004993127658963203, 0.03191784396767616, -0.08459635078907013, -0.13853681087493896, 0.004789369646459818, 0.33449316024780273, -0.11318263411521912, 0.013498755171895027, -0.025389941409230232, 0.053168006241321564, -0.05742674320936203, 0.08456550538539886, -0.07715325802564621, -0.04888878017663956, 0.09096968173980713, -0.21887867152690887, 0.03743908181786537, 0.07216470688581467, 0.017166929319500923, -0.07699387520551682, 0.0932677686214447, -0.11784102022647858, 0.027869997546076775, 0.1565779596567154, -0.3090413808822632, -0.048322949558496475, -0.05409501492977142, 0.06778387725353241, 0.1375984251499176, 0.044654034078121185, 0.10584048926830292, 0.08056485652923584, -0.03216809406876564, 0.09942833334207535, 0.015933426097035408, -0.14016255736351013, 0.0003787234891206026, 0.08693891018629074, 0.015894418582320213, -0.14782875776290894, 0.23973268270492554, 0.03688552603125572, -0.09386827051639557, -0.06660136580467224, 0.23404470086097717, -0.05190868303179741, -0.061600442975759506, -0.143992617726326, 0.08999405056238174, -0.06630992889404297, -0.12268120050430298, 0.040599625557661057, -0.01828654482960701, 0.03716189041733742, 0.21800193190574646, 0.04191570729017258, 0.17347969114780426, 0.03973954916000366, -0.05395336076617241, 0.1816711574792862, -0.011299784295260906, -0.08868960291147232, -0.04015281796455383, -0.09076227992773056, -0.1533331423997879, -0.041930459439754486, 0.26126015186309814, -0.13060596585273743, -0.09613201022148132, -0.11954604834318161, 0.11714274436235428, -0.11608145385980606, -0.07262169569730759, 0.006457416340708733, -0.06323787569999695, -0.0692286267876625, -0.11825434863567352, -0.035021789371967316, -0.0114168431609869, -0.11120990663766861, 0.08548320829868317, -0.000793378334492445, -0.0059143309481441975, -0.06473533064126968, 0.01710466854274273, 0.14210131764411926, 0.011864878237247467, 0.11346175521612167, 0.16409091651439667, 0.06108727678656578, 0.21627581119537354, -0.10105438530445099, -0.11361472308635712, 0.12454980611801147, -0.07703795284032822, -0.030023952946066856, 0.3381611406803131, -0.09178240597248077, -0.014071674086153507, -0.01994057558476925, 0.0648939236998558, -0.04699146747589111, -0.09501075744628906, 0.07164665311574936, -0.017618414014577866, -0.20475350320339203, 0.012440398335456848, -0.21029751002788544, 0.10059109330177307, 0.025866519659757614, 0.07438123971223831, 0.08806432783603668, 0.07154779136180878, 0.05169031023979187, 0.03326669707894325, 0.008451185189187527, -0.09226997941732407, -0.005910283885896206, -0.0027855176012963057, -0.022772042080760002, -0.07417235523462296, 0.31286412477493286, -0.1473950892686844, -0.09874339401721954, 0.03616202622652054, 0.10359495878219604, -0.09471222013235092, 0.014124877750873566, 0.15414123237133026, 0.15688680112361908, -0.045490559190511703, -0.25922855734825134, -0.0362209789454937, -0.04336727410554886, -0.09494815766811371, 0.144168421626091, 0.03821916505694389, 0.13758493959903717, 0.06886949390172958, 0.024475645273923874, -0.02738124132156372, -0.024884961545467377, -0.18191075325012207, -0.09152158349752426, -0.12596341967582703, 0.024104436859488487, 0.04982855170965195, 0.17658577859401703, 0.07094301283359528, -0.04141463339328766, -0.01361959520727396, 0.05363116413354874, -0.07422400265932083, -0.05049265921115875, 0.0038002405781298876, -0.1914263814687729, 0.04547438398003578, 0.02596084028482437, 0.07177992165088654, 0.22288240492343903, 0.0290970578789711, -0.06169590726494789, 0.07056407630443573, 0.06419803202152252, -0.1097598522901535, 0.07025780528783798, -0.023991432040929794, 0.06403990834951401, 0.010075516998767853, -0.01497514359652996, 0.021469060331583023, 0.01163057703524828, -0.06307860463857651, 0.0665537491440773, -0.012200938537716866, -0.07270172983407974, -0.1948794424533844, -0.11466846615076065, -0.04243917763233185, 0.14924412965774536, -0.10674240440130234, 0.018771111965179443, 0.023538850247859955, 0.06785940378904343, 0.0229969322681427, 0.14905935525894165, -0.018309470266103745, -0.011103630997240543, 0.06105777993798256, -0.06776928901672363, -0.052950579673051834, 0.2104865312576294, -0.0729382261633873, -0.06132755056023598, -0.028726842254400253, 0.21022574603557587, 0.25954172015190125, -0.0872897356748581, 0.04502170905470848, -0.019564826041460037, 0.06368174403905869, 0.08897557854652405, 0.07044430822134018, -0.03920239210128784, 0.277360200881958, -0.13055098056793213, -0.09941151738166809, 0.05294721573591232, 0.051166798919439316, -0.014584597200155258, 0.0040670037269592285, 0.0024921013973653316, -0.04536895081400871, -0.0929202288389206, 0.11709604412317276, -0.21066778898239136, 0.16287678480148315, -0.03428579121828079, -0.17375168204307556, -0.03228078782558441, -0.03540673479437828, 0.15313555300235748, 0.04490251839160919, 0.11070434004068375, -0.052801214158535004, -0.15846936404705048, -0.061232149600982666, 0.04336298257112503, -0.2800070643424988, -0.21539540588855743, 0.02419181913137436, 0.0668206438422203, 0.03457409888505936, -0.03128748759627342, 0.04788980260491371, -0.09045972675085068, 0.011143038049340248, -0.0347188338637352, 0.004606267903000116, 0.04971807822585106, -0.014385053887963295, -0.18549744784832, 0.05256281793117523, 0.041265591979026794, -0.09536474943161011, 0.1594448834657669, -0.10768021643161774, -0.035135116428136826, 0.04182901605963707, -0.03044901043176651, 0.059542424976825714, 0.0014320610789582133, -0.038819510489702225, 0.005382135976105928, 0.039671868085861206, -0.07253634184598923, -0.01786722056567669, -0.0003302812110632658, -0.001146028283983469, 0.05443781241774559, -0.09255039691925049, -0.07834245264530182, 0.05218031629920006, -0.0031077188905328512, 0.07994593679904938, 0.039965156465768814, 0.017863955348730087, 0.07039175927639008, -0.09229789674282074, 0.12315071374177933, 0.0015282738022506237, 0.05944240093231201, 0.1489715278148651, -0.03311542049050331, 0.039575956761837006, -0.30244919657707214, 0.13927175104618073, -0.03643213212490082, -0.07314969599246979, -0.12730808556079865 ]
null
null
transformers
# Eren Yeager DialoGPT Model
{"tags": ["conversational"]}
text-generation
Awsaf/DialoGPT-medium-eren
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Eren Yeager DialoGPT Model
[ "# Eren Yeager DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Eren Yeager DialoGPT Model" ]
[ 51, 10 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Eren Yeager DialoGPT Model" ]
[ -0.027911314740777016, 0.002212866209447384, -0.006088864989578724, 0.02654150314629078, 0.13750776648521423, -0.004181576427072287, 0.11697821319103241, 0.10612907260656357, -0.004773078486323357, -0.04564722627401352, 0.12719190120697021, 0.1579304337501526, -0.014811663888394833, 0.08041273057460785, -0.0664093866944313, -0.2947358191013336, 0.056030672043561935, 0.04018423706293106, 0.03296096622943878, 0.09873169660568237, 0.10458534210920334, -0.0374612919986248, 0.0901694968342781, 0.04182201623916626, -0.13502612709999084, 0.030983274802565575, -0.016522521153092384, -0.10926976054906845, 0.12694384157657623, 0.0746646597981453, 0.019994035363197327, 0.015364064835011959, -0.0342523530125618, -0.15600478649139404, 0.033812131732702255, -0.02927728183567524, -0.06262387335300446, 0.02605077065527439, -0.0005621216841973364, -0.11269515007734299, 0.09907890111207962, 0.08447273075580597, -0.009351576678454876, 0.025492960587143898, -0.14146776497364044, 0.001527918502688408, 0.01729406975209713, 0.06338765472173691, 0.05408499389886856, 0.11154033988714218, -0.05224013701081276, 0.09499555081129074, -0.07487846910953522, 0.10449241101741791, 0.08377719670534134, -0.3168579041957855, -0.005425960291177034, 0.0730980709195137, 0.027854446321725845, 0.028805045410990715, -0.0547286756336689, 0.05906819924712181, 0.042627736926078796, 0.017213687300682068, -0.03681536018848419, -0.07491102069616318, -0.05005268380045891, 0.019782904535531998, -0.09471537172794342, 0.009558717720210552, 0.23813758790493011, -0.029115241020917892, 0.07244504243135452, -0.07173268496990204, -0.0974801778793335, 0.011081937700510025, -0.025068316608667374, -0.03671465069055557, -0.08341493457555771, 0.06482242047786713, -0.0024508810602128506, -0.1064898818731308, -0.1324198991060257, -0.002216493943706155, -0.18776017427444458, 0.19996824860572815, 0.05922221392393112, 0.02688816748559475, -0.1968662142753601, 0.0871189683675766, -0.03405719995498657, -0.08502621948719025, 0.003933928906917572, -0.08732712268829346, 0.011266142129898071, 0.017307167872786522, -0.03054118901491165, -0.020367981866002083, 0.07375447452068329, 0.13741238415241241, -0.028591550886631012, 0.03023492358624935, 0.01960638165473938, 0.06410736590623856, 0.05705239996314049, 0.08472597599029541, 0.007407806813716888, -0.05856510251760483, 0.025626128539443016, -0.09717803448438644, 0.0044317529536783695, -0.04431717470288277, -0.19041424989700317, -0.029886845499277115, 0.06511785089969635, 0.03237251937389374, -0.0006512009422294796, 0.1151537224650383, -0.008167045190930367, -0.03870274871587753, 0.0639524832367897, -0.01447448879480362, -0.023046128451824188, 0.014797359704971313, -0.01271944958716631, 0.12411478161811829, 0.004054410848766565, 0.05897087976336479, -0.12118732929229736, 0.015976181253790855, -0.05778743699193001, -0.009067374281585217, -0.008505940437316895, -0.047254689037799835, -0.01196365337818861, -0.028562892228364944, 0.03950364142656326, -0.16678299009799957, -0.17535300552845, -0.007819981314241886, -0.014465922489762306, -0.05572028458118439, -0.09783224761486053, -0.14028844237327576, -0.02101544477045536, 0.04961876943707466, -0.06024191156029701, -0.029810359701514244, -0.06405014544725418, 0.056205712258815765, -0.038146600127220154, 0.07399924844503403, -0.10696495324373245, 0.08014174550771713, -0.0972856730222702, -0.029659582301974297, -0.061055947095155716, 0.12308347970247269, 0.018082154914736748, 0.04056892544031143, -0.010664043948054314, -0.03748271241784096, -0.09845724701881409, 0.07997757196426392, -0.02945392206311226, 0.23414942622184753, -0.06259091198444366, -0.12329970300197601, 0.2446746677160263, -0.06195869296789169, -0.10478131473064423, 0.13243615627288818, -0.011327220126986504, 0.07630936801433563, 0.12742087244987488, 0.23806971311569214, -0.0014718718593940139, -0.002021905966103077, 0.07660718262195587, 0.09772556275129318, -0.0888412669301033, 0.010308305732905865, 0.017693905159831047, -0.01694844849407673, -0.06788451969623566, 0.030050000175833702, 0.07075970619916916, 0.0671108216047287, -0.057543087750673294, -0.007923207245767117, 0.013506937772035599, 0.00465764245018363, 0.10144679248332977, -0.05060725286602974, 0.1264316886663437, -0.042693186551332474, -0.05461583286523819, 0.004241661634296179, 0.02134433202445507, -0.07027986645698547, 0.04334379732608795, -0.06212904304265976, 0.06464461982250214, 0.018748203292489052, 0.06633640080690384, -0.11670234054327011, -0.05321534723043442, -0.04361887276172638, 0.14034244418144226, 0.07160402834415436, 0.12365050613880157, 0.062282055616378784, -0.01821415312588215, -0.0138564957305789, 0.006879863794893026, 0.12376219034194946, -0.015102057717740536, -0.06797275692224503, -0.09283650666475296, 0.1066327691078186, -0.05727006494998932, 0.08200646936893463, -0.07870342582464218, 0.003197659971192479, 0.0071099428460001945, 0.11565310508012772, -0.015243450179696083, 0.04568905010819435, -0.016619978472590446, -0.01683552749454975, -0.05574291944503784, 0.023832296952605247, 0.08471713960170746, -0.0011261670151725411, -0.09090180695056915, 0.24301230907440186, -0.157989963889122, 0.13089437782764435, 0.16980338096618652, -0.22059747576713562, 0.000602465879637748, -0.13587455451488495, -0.04118923097848892, -0.002254024613648653, 0.0565236359834671, -0.056172795593738556, 0.29226094484329224, -0.013786539435386658, 0.18275363743305206, -0.04036833718419075, -0.053553253412246704, -0.029752690345048904, -0.06536604464054108, 0.0031888976227492094, 0.12355395406484604, 0.042509518563747406, -0.14467087388038635, 0.1534702479839325, 0.08020243793725967, 0.031901728361845016, 0.2069423645734787, 0.049485474824905396, 0.012171273119747639, 0.07210048288106918, -0.007238812278956175, -0.07268530875444412, -0.06759923696517944, -0.28957924246788025, -0.05222652107477188, 0.05851839482784271, 0.060140419751405716, 0.1279224008321762, -0.09632540494203568, -0.020418690517544746, -0.004157797899097204, -0.01677325740456581, 0.052272554486989975, 0.12042442709207535, 0.019698934629559517, 0.10366007685661316, -0.013060590252280235, -0.06272117048501968, 0.05990612506866455, 0.02485082857310772, -0.08760237693786621, 0.1871201992034912, -0.12305431813001633, -0.348667174577713, -0.11102789640426636, -0.16467857360839844, -0.06480559706687927, 0.04629170522093773, 0.11966551095247269, -0.12511925399303436, -0.02569965273141861, 0.007567602209746838, 0.15048475563526154, -0.09380864351987839, -0.0032671347726136446, -0.013771415688097477, -0.015438703820109367, -0.126917764544487, -0.10790715366601944, -0.05371473357081413, -0.052555907517671585, -0.04506136476993561, 0.12442152947187424, -0.1258419305086136, 0.03499980270862579, 0.18617478013038635, 0.05790780484676361, 0.05070700868964195, -0.039302438497543335, 0.24451181292533875, -0.10882314294576645, 0.014103914611041546, 0.16823236644268036, -0.05550368130207062, 0.07239317893981934, 0.1257844865322113, -0.02239038236439228, -0.06186822056770325, 0.0417042076587677, -0.01564779318869114, -0.05637405067682266, -0.20563599467277527, -0.14394646883010864, -0.10536088794469833, 0.06093553081154823, 0.016695957630872726, 0.047383975237607956, 0.12462984025478363, 0.06735464930534363, -0.038802698254585266, 0.010904701426625252, 0.0584658682346344, 0.08733946830034256, 0.30061331391334534, -0.05780338495969772, 0.13155458867549896, 0.0030236602760851383, -0.18689827620983124, 0.08390525728464127, 0.08054947853088379, 0.10681704431772232, 0.054452523589134216, 0.06009865179657936, 0.022396378219127655, 0.017471475526690483, 0.11721302568912506, 0.048886507749557495, 0.012268845923244953, -0.03947854042053223, -0.04356984421610832, -0.02080397866666317, 0.014941216446459293, 0.06450910121202469, 0.04430297389626503, -0.15766146779060364, -0.038146354258060455, 0.009479164145886898, 0.0888020470738411, 0.020039884373545647, 0.08759208023548126, -0.15497662127017975, -0.024069448933005333, 0.06510753184556961, -0.054713718593120575, -0.10448204725980759, 0.08614546060562134, 0.02486874908208847, -0.1321134865283966, 0.039483942091464996, 0.011853659525513649, 0.10699663311243057, -0.10868103057146072, 0.0862540453672409, -0.11710895597934723, -0.021183105185627937, 0.011831039562821388, 0.11326701939105988, -0.24599456787109375, 0.21163731813430786, -0.023929975926876068, -0.04800255596637726, -0.1052757278084755, -0.029741162434220314, 0.006329202558845282, 0.06953833252191544, 0.11575132608413696, -0.0012118960730731487, 0.03964276239275932, 0.015342460945248604, -0.03450242057442665, 0.0380200557410717, 0.1144038736820221, -0.029734879732131958, -0.01223069243133068, -0.03882564604282379, 0.005917138420045376, -0.0152976606041193, -0.05978953838348389, -0.016072135418653488, -0.1818811297416687, 0.04910050332546234, 0.07007049769163132, 0.09170157462358475, 0.041362129151821136, -0.04407643526792526, -0.10815072059631348, 0.23942065238952637, -0.008049231953918934, -0.10337239503860474, -0.08201788365840912, -0.032130010426044464, 0.04759381711483002, -0.08160634338855743, 0.02442564070224762, -0.05011758208274841, 0.03462117165327072, -0.07819290459156036, -0.16515016555786133, 0.09629511088132858, -0.09478511661291122, -0.04210911691188812, -0.004007741343230009, 0.20273327827453613, 0.009058685973286629, 0.016706295311450958, 0.05353778973221779, 0.0016211394686251879, -0.1425393968820572, -0.11372096836566925, 0.0013482157373800874, -0.003324324730783701, -0.007970394566655159, 0.025335948914289474, -0.0337115079164505, -0.07382581382989883, -0.06032861769199371, -0.03249669075012207, 0.3192692697048187, 0.1717197149991989, -0.030084926635026932, 0.14904461801052094, 0.1356353461742401, -0.04841040447354317, -0.2829614579677582, -0.12140721082687378, -0.057287342846393585, -0.01086074486374855, -0.09970887005329132, -0.17393837869167328, 0.08232548087835312, -0.005169311072677374, -0.0005034233327023685, 0.08601309359073639, -0.27204909920692444, -0.1208847388625145, 0.17322105169296265, -0.037148844450712204, 0.4350094199180603, -0.05628301575779915, -0.07903409004211426, -0.047017235308885574, -0.1264137178659439, 0.13038873672485352, -0.027262581512331963, 0.11540257930755615, -0.005956257693469524, 0.18229138851165771, 0.036609284579753876, 0.009509388357400894, 0.08200092613697052, 0.06011057272553444, -0.04508582502603531, -0.07879867404699326, -0.06836969405412674, -0.0409734845161438, 0.03996657207608223, 0.059003666043281555, -0.05674707144498825, 0.030272098258137703, -0.0912083238363266, -0.06687659025192261, -0.08868037164211273, 0.0499163419008255, 0.04004964977502823, -0.07947522401809692, -0.005333843175321817, -0.05748020112514496, -0.008559823967516422, 0.010271214880049229, 0.16289633512496948, -0.07863648980855942, 0.09052719175815582, 0.05403773859143257, 0.11378531903028488, -0.13655376434326172, -0.01898067630827427, -0.05986408516764641, -0.058018606156110764, 0.058876000344753265, -0.06683717668056488, 0.020161481574177742, 0.12210091203451157, -0.03988698497414589, 0.10270705074071884, 0.09335129708051682, -0.013463820330798626, 0.034823719412088394, 0.09486859291791916, -0.24312429130077362, -0.08076094835996628, -0.0900893285870552, -0.03174050152301788, 0.10241445153951645, 0.1064629852771759, 0.22250035405158997, -0.014801852405071259, -0.03513466939330101, -0.004578572232276201, 0.014889140613377094, -0.045643724501132965, 0.0709235742688179, 0.013484841212630272, 0.0069953990168869495, -0.14106491208076477, 0.060476161539554596, -0.010529408231377602, -0.08200801908969879, 0.05043152719736099, 0.1456608921289444, -0.11437234282493591, -0.11853617429733276, -0.11360252648591995, 0.14504653215408325, -0.16701167821884155, -0.021376926451921463, -0.04162393882870674, -0.12567441165447235, 0.06344221532344818, 0.07019254565238953, 0.060600534081459045, 0.06011160463094711, -0.07699717581272125, -0.012427572160959244, -0.0038963707629591227, 0.0008886018767952919, 0.07966440916061401, -0.02905196323990822, -0.06216762959957123, 0.02586660161614418, -0.03520284965634346, 0.1291937381029129, -0.09012147784233093, -0.09590278565883636, -0.1565147489309311, 0.02994227409362793, -0.10667969286441803, -0.06965557485818863, -0.10032203793525696, -0.0428440198302269, -0.00262360367923975, -0.045926377177238464, -0.03227584436535835, -0.025315774604678154, -0.10138493776321411, 0.03240727633237839, -0.030576108023524284, 0.016016555950045586, -0.08088994771242142, 0.02305862493813038, 0.06347178667783737, -0.02922985702753067, 0.15837180614471436, 0.1372847855091095, -0.11251524835824966, 0.09979286044836044, -0.14255216717720032, -0.07231959700584412, 0.11594482511281967, 0.019655069336295128, 0.06286979466676712, 0.06191590428352356, -0.006000985391438007, 0.06017283722758293, 0.06621479988098145, 0.04653143882751465, 0.03415650501847267, -0.08326822519302368, 0.04200566187500954, -0.0646275132894516, -0.1467646360397339, -0.04541188105940819, -0.03522748500108719, -0.00006641871732426807, 0.049913663417100906, 0.10468986630439758, -0.06630338728427887, 0.07978479564189911, -0.03335186839103699, 0.0485500693321228, 0.04283009469509125, -0.16071493923664093, -0.0016767423367127776, -0.09866170585155487, 0.0401577390730381, 0.006657737772911787, 0.2117137461900711, 0.013684890232980251, 0.0027913604862987995, 0.024495098739862442, 0.0736219733953476, -0.02743598073720932, -0.002471407176926732, 0.15180867910385132, 0.08780793100595474, -0.04530421271920204, -0.13664647936820984, 0.08738385140895844, 0.0405094213783741, 0.01631191000342369, 0.10383269935846329, -0.031829915940761566, -0.008556324057281017, 0.09780613332986832, -0.0038412895519286394, 0.015561237931251526, -0.15628916025161743, -0.1354246437549591, -0.02246163599193096, 0.04791790992021561, -0.04586232081055641, 0.12715955078601837, 0.1471618115901947, -0.02493971586227417, 0.018831228837370872, 0.00042010590550489724, -0.0704737976193428, -0.16401414573192596, -0.17520754039287567, -0.09627661854028702, -0.14643602073192596, 0.0172544214874506, -0.13622133433818817, 0.021934010088443756, 0.003999766428023577, 0.0960099846124649, -0.0583292618393898, 0.09210488200187683, 0.04854535311460495, -0.13385102152824402, 0.07388214021921158, -0.03764290362596512, 0.09274911880493164, -0.01758461631834507, -0.0007006488158367574, -0.07543523609638214, 0.04523608833551407, 0.004211790394037962, 0.042417898774147034, -0.04101572185754776, -0.0046655829064548016, -0.12902702391147614, -0.06667015701532364, -0.06838233023881912, 0.08034592121839523, -0.01676604337990284, 0.11996539682149887, 0.019604148343205452, -0.03948596864938736, 0.01804964430630207, 0.23501405119895935, -0.05713358148932457, -0.11148800700902939, -0.06981673836708069, 0.18067556619644165, 0.04537436366081238, 0.1319604516029358, -0.015881136059761047, -0.024500006809830666, -0.09240072220563889, 0.3439176380634308, 0.29867371916770935, -0.09610675275325775, 0.016190994530916214, 0.014121510088443756, 0.04901411756873131, 0.1302851140499115, 0.06287132203578949, 0.12715020775794983, 0.29735445976257324, -0.07013385742902756, -0.03710533678531647, -0.0347326435148716, -0.02461455762386322, -0.0571853369474411, 0.049839600920677185, 0.049518048763275146, -0.06330085545778275, -0.0428764782845974, 0.09395898878574371, -0.2689298391342163, 0.06985916942358017, -0.18323922157287598, -0.16338320076465607, -0.07951278239488602, 0.005766971968114376, 0.10535693168640137, 0.04539991170167923, 0.11296820640563965, 0.0022724000737071037, -0.08306769281625748, 0.06235331669449806, 0.02735518477857113, -0.19074445962905884, 0.025357622653245926, 0.07505623251199722, -0.03211485967040062, -0.054898057132959366, -0.02451353147625923, 0.09864864498376846, 0.07073848694562912, 0.06140178069472313, 0.003152213990688324, 0.0504014827311039, -0.014957930892705917, -0.0376725010573864, 0.027445264160633087, 0.038267020136117935, 0.027152271941304207, -0.11280703544616699, 0.07820769399404526, -0.17490987479686737, 0.03504842147231102, 0.05765896663069725, -0.038286130875349045, -0.013074805960059166, 0.05193670839071274, -0.07573163509368896, 0.0699928030371666, 0.07658235728740692, 0.0068557532504200935, -0.023523109033703804, -0.021797090768814087, -0.013556537218391895, -0.02397061698138714, -0.07137176394462585, -0.07505928725004196, -0.1653105765581131, -0.10286615043878555, 0.09358075261116028, 0.00019320697174407542, -0.16916175186634064, 0.0002449512539897114, -0.13498690724372864, 0.07321856915950775, -0.11721865832805634, 0.1098661795258522, 0.08218306303024292, 0.023247461766004562, -0.0023787375539541245, 0.025714147835969925, 0.03436719626188278, 0.06298618018627167, -0.10391530394554138, -0.10047544538974762 ]
null
null
transformers
# Eren Yeager Model
{"tags": ["conversational"]}
text-generation
Awsaf/large-eren
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Eren Yeager Model
[ "# Eren Yeager Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Eren Yeager Model" ]
[ 51, 6 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Eren Yeager Model" ]
[ -0.025332892313599586, -0.04147571697831154, -0.005935561843216419, 0.041260406374931335, 0.1450556218624115, 0.00268109911121428, 0.08883140236139297, 0.09087996184825897, -0.008285000920295715, -0.03139584884047508, 0.16023309528827667, 0.1501762419939041, -0.03001871332526207, 0.05496054142713547, -0.06282716244459152, -0.28381142020225525, 0.057411398738622665, 0.02947804518043995, 0.01682799495756626, 0.08504635095596313, 0.09764163196086884, -0.055366937071084976, 0.11479885131120682, 0.05385094881057739, -0.1288566291332245, 0.03836223855614662, -0.038500234484672546, -0.09769318252801895, 0.1270838975906372, 0.0801425352692604, 0.028572576120495796, 0.020089836791157722, -0.014049501158297062, -0.14782948791980743, 0.035911720246076584, -0.03042929247021675, -0.08472075313329697, 0.029137540608644485, 0.015061143785715103, -0.13441303372383118, 0.05650213733315468, 0.06247495859861374, -0.022475406527519226, 0.011940686963498592, -0.13955236971378326, 0.040894076228141785, 0.02930375002324581, 0.039102740585803986, 0.048634979873895645, 0.12122365087270737, -0.04538392275571823, 0.0709574744105339, -0.08175098150968552, 0.09074386954307556, 0.1207471415400505, -0.28602463006973267, 0.0022194853518158197, 0.056873925030231476, 0.03753066435456276, 0.007650463376194239, -0.06049879267811775, 0.04083240404725075, 0.06617356091737747, 0.024726076051592827, -0.004049189854413271, -0.06826625764369965, -0.0005206406931392848, 0.05364646017551422, -0.09522785246372223, 0.004577619954943657, 0.2210826277732849, -0.03577026352286339, 0.0740099847316742, -0.04828786477446556, -0.11615590751171112, 0.031207837164402008, -0.014034699648618698, -0.013528143987059593, -0.06901992857456207, 0.05525771901011467, -0.0000323316126014106, -0.09702093154191971, -0.13282625377178192, -0.0029844543896615505, -0.18606695532798767, 0.22660478949546814, 0.06375082582235336, 0.026750560849905014, -0.19960571825504303, 0.07902806997299194, -0.030362050980329514, -0.08079653233289719, 0.006133316084742546, -0.07882731407880783, 0.03248396888375282, 0.01580384373664856, -0.05448947846889496, -0.02512873336672783, 0.08831696957349777, 0.14854884147644043, -0.0313287116587162, 0.028517764061689377, 0.07729803770780563, 0.07578308880329132, 0.07601812481880188, 0.08174417167901993, 0.020381024107336998, -0.039386678487062454, 0.027027368545532227, -0.08788836002349854, 0.005306021310389042, -0.048079006373882294, -0.1927737593650818, -0.01623283326625824, 0.0684148445725441, 0.013770068064332008, -0.034688327461481094, 0.09830816090106964, -0.0127619793638587, -0.037294819951057434, 0.06878458708524704, -0.018691962584853172, -0.028169695287942886, 0.014599401503801346, -0.012463817372918129, 0.10653620958328247, -0.019575195387005806, 0.05872434377670288, -0.10471490770578384, 0.04302217811346054, -0.0729769691824913, -0.021136244758963585, -0.016918040812015533, -0.0617886446416378, -0.02200259454548359, -0.041385117918252945, 0.07140568643808365, -0.16090045869350433, -0.21170975267887115, -0.014553609304130077, -0.006137639284133911, -0.05227511748671532, -0.08516639471054077, -0.14034618437290192, -0.024745475500822067, 0.042744703590869904, -0.05039564147591591, -0.005774127319455147, -0.07784910500049591, 0.03866227716207504, -0.05145882070064545, 0.053758300840854645, -0.10314833372831345, 0.0864669680595398, -0.10810431838035583, -0.013005821034312248, -0.02150745503604412, 0.10055819153785706, 0.03490517660975456, 0.05044754967093468, 0.002909333212301135, -0.050482068210840225, -0.0882713720202446, 0.09510049968957901, -0.023735130205750465, 0.21454717218875885, -0.05468469858169556, -0.14443252980709076, 0.18999706208705902, -0.06611760705709457, -0.11318111419677734, 0.11519227921962738, -0.028126724064350128, 0.05073579028248787, 0.10728960484266281, 0.2441273331642151, -0.02093513496220112, 0.01960749179124832, 0.09295392781496048, 0.08085092902183533, -0.09640238434076309, 0.015445905737578869, 0.015160854905843735, -0.0009533722186461091, -0.09181521832942963, 0.028442852199077606, 0.05489760637283325, 0.05226320028305054, -0.07410843670368195, 0.008690490387380123, 0.014048758894205093, 0.00358353229239583, 0.09059233218431473, -0.05539718270301819, 0.13503322005271912, -0.052298810333013535, -0.03768755868077278, -0.006264116149395704, 0.012195298448204994, -0.08072612434625626, 0.03793025389313698, -0.033451348543167114, 0.09664286673069, 0.024308722466230392, 0.06400270015001297, -0.128214031457901, -0.04907882586121559, -0.033124033361673355, 0.0970255434513092, 0.05423706769943237, 0.136301189661026, 0.057435404509305954, 0.017302164807915688, 0.003600876312702894, -0.02169531211256981, 0.09526179730892181, -0.0061325752176344395, -0.07655175775289536, -0.09369024634361267, 0.0853009819984436, -0.05962945893406868, 0.036766938865184784, -0.10916823148727417, 0.0060860044322907925, 0.01314573548734188, 0.10830754786729813, -0.015849456191062927, 0.0704508125782013, -0.042032063007354736, -0.022903628647327423, -0.07464050501585007, 0.03333943709731102, 0.08244634419679642, -0.0046594212763011456, -0.09943842142820358, 0.21786871552467346, -0.15171940624713898, 0.15467774868011475, 0.1641250103712082, -0.24636009335517883, 0.009679614566266537, -0.1187979206442833, -0.05155986547470093, -0.002095184288918972, 0.02084474079310894, -0.06590195000171661, 0.30824005603790283, -0.0056832595728337765, 0.19233836233615875, -0.0420638844370842, -0.06623262912034988, -0.017777057364583015, -0.05956543609499931, 0.005848335567861795, 0.1461973786354065, 0.021163366734981537, -0.1439790576696396, 0.1555645614862442, 0.10730569809675217, 0.009294564835727215, 0.174089714884758, 0.05619002878665924, 0.007433631457388401, 0.08914954960346222, -0.027163896709680557, -0.07943092286586761, -0.05232762545347214, -0.24356670677661896, -0.06253110617399216, 0.04718714952468872, 0.053997885435819626, 0.12697887420654297, -0.1078331395983696, -0.017801349982619286, 0.0010772977257147431, -0.0015619826735928655, 0.049308713525533676, 0.10751854628324509, 0.015203947201371193, 0.08618606626987457, -0.013338983990252018, -0.04502520710229874, 0.0605718195438385, 0.02989453636109829, -0.08777331560850143, 0.20137900114059448, -0.09335756301879883, -0.31888654828071594, -0.12773975729942322, -0.14392408728599548, -0.07132304459810257, 0.046905145049095154, 0.12775051593780518, -0.11761951446533203, -0.032858870923519135, -0.006228969432413578, 0.17725156247615814, -0.1039329245686531, -0.006700340658426285, -0.007657581474632025, -0.00777060491964221, -0.12006410211324692, -0.09683465957641602, -0.05599261075258255, -0.06419773399829865, -0.013667027465999126, 0.1365424543619156, -0.11995524913072586, 0.060100048780441284, 0.13404114544391632, 0.0599537268280983, 0.054190512746572495, -0.033132027834653854, 0.28655293583869934, -0.0842575952410698, 0.018927454948425293, 0.16770215332508087, -0.06689518690109253, 0.09282271564006805, 0.10691256076097488, -0.019131751731038094, -0.06777849793434143, 0.0445442833006382, -0.0027632294222712517, -0.06057547777891159, -0.19518323242664337, -0.12939679622650146, -0.10115288943052292, 0.043114326894283295, 0.04112556576728821, 0.05564709007740021, 0.09474273025989532, 0.07836392521858215, -0.022328514605760574, 0.04697984457015991, 0.04529176279902458, 0.0962146669626236, 0.2840481698513031, -0.029413405805826187, 0.12374106049537659, 0.01896473579108715, -0.1975119560956955, 0.09516526013612747, 0.08875338733196259, 0.1518971025943756, 0.05294444411993027, 0.0774402916431427, 0.023680537939071655, -0.0034323676954954863, 0.12368348985910416, 0.06475471705198288, -0.0006220728973858058, -0.027110718190670013, -0.05925340950489044, -0.00321245682425797, 0.03436891362071037, 0.07888416945934296, 0.007087775971740484, -0.1430618315935135, -0.06446491181850433, 0.025035493075847626, 0.10551818460226059, 0.04173598438501358, 0.07679212093353271, -0.1893318146467209, -0.030849482864141464, 0.0769379734992981, -0.06567714363336563, -0.08370593935251236, 0.08808514475822449, 0.009801344014704227, -0.1267211139202118, 0.03315023332834244, 0.005362953525036573, 0.1251474767923355, -0.10900801420211792, 0.10789594799280167, -0.09420333057641983, 0.011182267218828201, 0.022115638479590416, 0.11585226655006409, -0.21780572831630707, 0.22621414065361023, -0.0322672463953495, -0.05440815910696983, -0.10187429934740067, -0.040575262159109116, 0.016846483573317528, 0.05632014200091362, 0.12387487292289734, 0.0030995216220617294, -0.005666235461831093, -0.012557543814182281, -0.003909221850335598, 0.05235520377755165, 0.12948237359523773, -0.07003723829984665, 0.007879682816565037, -0.043171897530555725, 0.01579933986067772, -0.022043989971280098, -0.019420228898525238, -0.016312196850776672, -0.1664101779460907, 0.03167984262108803, 0.039695873856544495, 0.08906092494726181, 0.03824755921959877, -0.04784121736884117, -0.11906388401985168, 0.20926479995250702, -0.035286519676446915, -0.09969821572303772, -0.09495671838521957, -0.007852165028452873, 0.04326080158352852, -0.08894498646259308, 0.014759697951376438, -0.051583901047706604, 0.05489485710859299, -0.07939047366380692, -0.17480474710464478, 0.09332851320505142, -0.09494546800851822, -0.06234513595700264, -0.0011284793727099895, 0.17227530479431152, 0.010835080407559872, 0.013674616813659668, 0.045076217502355576, 0.018028752878308296, -0.18299756944179535, -0.12475068122148514, 0.008418966084718704, -0.03295714408159256, 0.025134654715657234, 0.012244915589690208, -0.048632144927978516, -0.029481129720807076, -0.04548950120806694, -0.029453841969370842, 0.3015056848526001, 0.18290795385837555, -0.023442037403583527, 0.13281695544719696, 0.13745518028736115, -0.04817730560898781, -0.2924574911594391, -0.1305389255285263, -0.07227805256843567, 0.013723672367632389, -0.08851484209299088, -0.17388297617435455, 0.09603485465049744, 0.02334379404783249, 0.010747838765382767, 0.0661388486623764, -0.2344924360513687, -0.12557357549667358, 0.15053188800811768, -0.010975034907460213, 0.4512232840061188, -0.05037682503461838, -0.07362699508666992, -0.0572483092546463, -0.09396239370107651, 0.10689058154821396, -0.08117593824863434, 0.11763514578342438, -0.012493251822888851, 0.17752650380134583, 0.022285427898168564, 0.0017594430828467011, 0.08271953463554382, 0.08173162490129471, -0.022780708968639374, -0.061397287994623184, -0.0748351514339447, -0.032503850758075714, 0.03499967232346535, 0.0521661713719368, -0.08570048213005066, 0.04901549965143204, -0.060450229793787, -0.06797277927398682, -0.0740925520658493, 0.04742326959967613, 0.04319363832473755, -0.08024751394987106, -0.0058230250142514706, -0.0649445652961731, 0.0022666258737444878, 0.013910967856645584, 0.19299329817295074, -0.04536239802837372, 0.055813275277614594, 0.03075716271996498, 0.09453325718641281, -0.15122584998607635, -0.029574835672974586, -0.05161730945110321, -0.04478799179196358, 0.0668744370341301, -0.0801779255270958, 0.02951093763113022, 0.14466986060142517, -0.04560340940952301, 0.10877559334039688, 0.11646720767021179, -0.003233896801248193, 0.03880559653043747, 0.08513949066400528, -0.25515562295913696, -0.07598037272691727, -0.10008913278579712, -0.05132655054330826, 0.10126889497041702, 0.10174956172704697, 0.21657487750053406, -0.0166829414665699, -0.03646595776081085, -0.012411174364387989, 0.005841774865984917, -0.03301382064819336, 0.053629495203495026, 0.025049487128853798, 0.012021643109619617, -0.13465255498886108, 0.061605989933013916, -0.017847323790192604, -0.0965965986251831, 0.0445428192615509, 0.13913311064243317, -0.12633934617042542, -0.11714956909418106, -0.09669703990221024, 0.20282451808452606, -0.19777968525886536, -0.028562666848301888, -0.06391964852809906, -0.1435028463602066, 0.07556240260601044, 0.066673144698143, 0.08305367827415466, 0.05390847101807594, -0.061368413269519806, -0.0022503710351884365, 0.005577721167355776, 0.002530223922803998, 0.09697824716567993, -0.028090234845876694, -0.06581524014472961, -0.03748657554388046, -0.02756102755665779, 0.14264020323753357, -0.09188636392354965, -0.07937917858362198, -0.15676762163639069, 0.024414489045739174, -0.12336313724517822, -0.04117114096879959, -0.10248744487762451, -0.04139011353254318, -0.006091339513659477, -0.04275057837367058, -0.03416808694601059, -0.016817796975374222, -0.09295547008514404, 0.024919109418988228, -0.03496759012341499, 0.026287874206900597, -0.07584398239850998, 0.012310133315622807, 0.0687403753399849, -0.04767940193414688, 0.14433103799819946, 0.12651380896568298, -0.11351826041936874, 0.0947682112455368, -0.13501513004302979, -0.0786503478884697, 0.10412468016147614, 0.019061120226979256, 0.07705588638782501, 0.05181018263101578, -0.011682084761559963, 0.048121750354766846, 0.041614532470703125, 0.031872864812612534, 0.018213963136076927, -0.07228616625070572, 0.038568392395973206, -0.056207627058029175, -0.15912435948848724, -0.052085403352975845, -0.056940868496894836, -0.006293867249041796, 0.05060161277651787, 0.12270944565534592, -0.05639200285077095, 0.08947080373764038, -0.04206082597374916, 0.054221734404563904, 0.059614066034555435, -0.1790173351764679, -0.05389028042554855, -0.11414217203855515, 0.028393026441335678, -0.009538494050502777, 0.24751603603363037, 0.0054687438532710075, 0.06008276715874672, 0.033084623515605927, 0.11718057096004486, -0.05012405291199684, 0.018808215856552124, 0.13949714601039886, 0.06486168503761292, -0.042782124131917953, -0.1718924641609192, 0.08123791962862015, 0.057473357766866684, 0.000540935609024018, 0.11073901504278183, -0.018669940531253815, 0.0064972927793860435, 0.10559894144535065, -0.007427590899169445, 0.023864813148975372, -0.1713230013847351, -0.09656883776187897, -0.00520862452685833, 0.047241248190402985, -0.007568977773189545, 0.11117103695869446, 0.15480712056159973, -0.03762107715010643, 0.012728477828204632, -0.00965136382728815, -0.06882303208112717, -0.1429436206817627, -0.17532972991466522, -0.11723683774471283, -0.14033518731594086, 0.017224596813321114, -0.14344565570354462, -0.005848214495927095, 0.017561331391334534, 0.08201192319393158, -0.04275894537568092, 0.08922991156578064, 0.07606778293848038, -0.12068917602300644, 0.05847769230604172, -0.04228857904672623, 0.0763988271355629, 0.003992842976003885, -0.013644612394273281, -0.08348160237073898, 0.024681787937879562, -0.018785906955599785, 0.03907422721385956, -0.020815076306462288, 0.0011063265847042203, -0.13467690348625183, -0.07356607913970947, -0.0676189512014389, 0.07636149972677231, -0.054904505610466, 0.06771336495876312, 0.014257930219173431, -0.04115178436040878, 0.021923931315541267, 0.2180529087781906, -0.04393543303012848, -0.07899454981088638, -0.06344340741634369, 0.15298046171665192, 0.0791470929980278, 0.13978630304336548, 0.01052096951752901, -0.02669629268348217, -0.09144168347120285, 0.35615336894989014, 0.30018532276153564, -0.12348666042089462, 0.01145854126662016, 0.0218402910977602, 0.04269815608859062, 0.1570618450641632, 0.02602466568350792, 0.11538645625114441, 0.256288081407547, -0.08990524709224701, -0.05645959824323654, -0.058366503566503525, -0.025968637317419052, -0.05392156168818474, 0.08465299755334854, 0.06255806237459183, -0.06454478204250336, -0.058805786073207855, 0.08028547465801239, -0.27832886576652527, 0.06424041092395782, -0.16518747806549072, -0.15986427664756775, -0.0730210393667221, 0.009001081809401512, 0.1136193796992302, 0.04140551760792732, 0.12754999101161957, -0.01110578328371048, -0.08936507999897003, 0.036556463688611984, 0.030792158097028732, -0.18954262137413025, 0.05358124151825905, 0.07636646926403046, -0.013592624105513096, -0.04615544527769089, -0.02707056514918804, 0.09311018139123917, 0.07119040191173553, 0.06440761685371399, -0.0005395222106017172, 0.05520762875676155, -0.016862833872437477, -0.0297293309122324, 0.014756396412849426, 0.030185414478182793, 0.02832975797355175, -0.14253491163253784, 0.07094985246658325, -0.2123117446899414, 0.0294964462518692, 0.09533986449241638, -0.04720396548509598, 0.013941444456577301, 0.043665722012519836, -0.06831807643175125, 0.06322167068719864, 0.06887345761060715, 0.01829863153398037, -0.028749212622642517, -0.026720790192484856, -0.005297311116009951, -0.022052567452192307, -0.06915225088596344, -0.06529081612825394, -0.15410590171813965, -0.09521839022636414, 0.10871907323598862, -0.020432719960808754, -0.18507342040538788, -0.030836261808872223, -0.1329219788312912, 0.08840413391590118, -0.11980156600475311, 0.10371612757444382, 0.10207953304052353, 0.019252965226769447, -0.017333662137389183, 0.058528367429971695, 0.025491751730442047, 0.05099969357252121, -0.07896925508975983, -0.12660005688667297 ]
null
null
transformers
# Rick DialoGPT Model
{"tags": ["conversational"]}
text-generation
Axcel/DialoGPT-small-rick
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Rick DialoGPT Model
[ "# Rick DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Rick DialoGPT Model" ]
[ 51, 7 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Rick DialoGPT Model" ]
[ -0.027243174612522125, 0.09208611398935318, -0.005486058536916971, 0.01197603065520525, 0.13312271237373352, -0.0006643096567131579, 0.14875547587871552, 0.13561291992664337, -0.012389403767883778, -0.048079900443553925, 0.13848258554935455, 0.20838283002376556, -0.007769247982650995, 0.06212212145328522, -0.07722679525613785, -0.3253750503063202, 0.05440690368413925, 0.05986349284648895, -0.02559526450932026, 0.11941008269786835, 0.10155656188726425, -0.034638021141290665, 0.07502283155918121, 0.008745936676859856, -0.1460564285516739, 0.011253442615270615, 0.020986590534448624, -0.11265120655298233, 0.11301227658987045, 0.0699501633644104, 0.03311868757009506, 0.044131726026535034, -0.04560676962137222, -0.12763948738574982, 0.04502782225608826, 0.00030866602901369333, -0.04332113638520241, 0.05997459217905998, 0.016281595453619957, -0.09000954777002335, 0.11693226546049118, 0.12603440880775452, -0.01263172086328268, 0.041781701147556305, -0.1548357903957367, -0.004369331523776054, -0.01233562733978033, 0.06789606809616089, 0.06087101250886917, 0.10755407065153122, -0.04065045714378357, 0.11729123443365097, -0.06241777911782265, 0.11526333540678024, 0.1129850223660469, -0.291816771030426, -0.016308816149830818, 0.14326390624046326, 0.043570004403591156, 0.04201141744852066, -0.04241296648979187, 0.09895236790180206, 0.01734745316207409, -0.009189855307340622, -0.04667704179883003, -0.07920589298009872, -0.0809992179274559, 0.022899743169546127, -0.08393258601427078, -0.009693359956145287, 0.24909301102161407, -0.033697742968797684, 0.07867740839719772, -0.07909003645181656, -0.08747624605894089, -0.011933685280382633, -0.03604159876704216, -0.03430533409118652, -0.10349667817354202, 0.07883962988853455, -0.03785189241170883, -0.09532928466796875, -0.11454451829195023, -0.029063701629638672, -0.16551746428012848, 0.1769428551197052, 0.028738701716065407, 0.03337583318352699, -0.22648879885673523, 0.09508261829614639, -0.012410550378262997, -0.09879330545663834, 0.018604513257741928, -0.08811058849096298, 0.012304049916565418, 0.017966609448194504, -0.025972042232751846, -0.002111254259943962, 0.08367783576250076, 0.11593183130025864, 0.01627914048731327, 0.018418017774820328, -0.01303142961114645, 0.05024925619363785, 0.039101485162973404, 0.07016518712043762, -0.018131986260414124, -0.026958800852298737, 0.025394905358552933, -0.09519384801387787, -0.01311302836984396, -0.06533002108335495, -0.19878731667995453, -0.008748088963329792, 0.05362382158637047, 0.059645626693964005, 0.040223345160484314, 0.1349429488182068, 0.005914759822189808, -0.04811347648501396, 0.041568055748939514, -0.017372997477650642, -0.016568226739764214, 0.013325352221727371, 0.004558354616165161, 0.14832930266857147, 0.012210249900817871, 0.05107790604233742, -0.11448643356561661, 0.0074756252579391, -0.04443434625864029, -0.019875049591064453, -0.033431850373744965, -0.05190093815326691, -0.010580608621239662, -0.024629589170217514, 0.015543424524366856, -0.1382266879081726, -0.1671048104763031, -0.0113193579018116, -0.006982414051890373, -0.04376089945435524, -0.11932645738124847, -0.1048901304602623, -0.03145192563533783, 0.04379252344369888, -0.060927584767341614, -0.0003760824038181454, -0.04660411551594734, 0.09378229826688766, -0.03543102741241455, 0.07682112604379654, -0.10023638606071472, 0.0828537717461586, -0.07001189142465591, -0.04422231763601303, -0.0734889879822731, 0.13164658844470978, 0.014363138936460018, 0.05487450957298279, -0.031934577971696854, -0.01827416382730007, -0.10224048048257828, 0.07911752909421921, -0.04339373856782913, 0.23623128235340118, -0.09449771791696548, -0.10362883657217026, 0.26979705691337585, -0.053989510983228683, -0.1375254988670349, 0.10795111209154129, -0.015854641795158386, 0.11475867033004761, 0.12686948478221893, 0.18240338563919067, 0.06434911489486694, 0.007867260836064816, 0.07431085407733917, 0.11333738267421722, -0.0774611383676529, -0.018117602914571762, 0.014873803593218327, -0.020292608067393303, -0.07848027348518372, 0.023533256724476814, 0.07671299576759338, 0.05307117849588394, -0.05429181456565857, -0.015286878682672977, 0.00432937266305089, 0.004517627414315939, 0.05698307976126671, -0.02530503273010254, 0.12313884496688843, -0.029461434110999107, -0.07295558601617813, -0.029503753408789635, 0.027530280873179436, -0.05828499048948288, 0.03278997913002968, -0.08230485767126083, 0.03637091815471649, -0.014406797476112843, 0.07024850696325302, -0.16572508215904236, -0.09323301911354065, -0.05250932276248932, 0.1899155229330063, 0.06807822734117508, 0.11413464695215225, 0.05567482113838196, -0.06841246038675308, -0.0038719952572137117, 0.018287649378180504, 0.1991138458251953, -0.01677977479994297, -0.07748494297266006, -0.09769339859485626, 0.10122697055339813, -0.07130109518766403, 0.06141059845685959, -0.050490207970142365, 0.017946461215615273, 0.020556224510073662, 0.1050461083650589, -0.03456922993063927, 0.039413414895534515, 0.011159577406942844, -0.034563858062028885, -0.06218598783016205, -0.004433273337781429, 0.09716981649398804, 0.0021626276429742575, -0.10631977766752243, 0.24286337196826935, -0.19168923795223236, 0.12176351994276047, 0.17641966044902802, -0.19923987984657288, -0.0002552573860157281, -0.11963175982236862, -0.026344671845436096, 0.011637656949460506, 0.037626978009939194, -0.042151857167482376, 0.24314165115356445, -0.00910688005387783, 0.16631373763084412, -0.03389734402298927, -0.04332707077264786, -0.041059546172618866, -0.046011339873075485, 0.010055569931864738, 0.11430004984140396, 0.1047205775976181, -0.17159950733184814, 0.17967921495437622, 0.05867021903395653, 0.05177219957113266, 0.16841758787631989, 0.018001655116677284, 0.021052619442343712, 0.06948674470186234, -0.003431870136409998, -0.03584783151745796, -0.07413756102323532, -0.2106374204158783, -0.023212855681777, 0.0793403834104538, 0.048357341438531876, 0.1068209707736969, -0.1037900522351265, -0.03368109092116356, -0.010547412559390068, -0.021230356767773628, 0.03035620041191578, 0.14086326956748962, 0.013085569255053997, 0.1286563277244568, -0.024180158972740173, -0.06866493821144104, 0.06965550780296326, 0.014881031587719917, -0.08571527898311615, 0.19352088868618011, -0.10702410340309143, -0.34334462881088257, -0.10363983362913132, -0.18596062064170837, -0.056601256132125854, 0.04553624242544174, 0.11461924016475677, -0.14119702577590942, -0.020731983706355095, 0.006813736632466316, 0.06912991404533386, -0.11165751516819, 0.01017086487263441, -0.03630850836634636, -0.017619650810956955, -0.13406261801719666, -0.1034051924943924, -0.05356309190392494, -0.044913630932569504, -0.05510649085044861, 0.12040390819311142, -0.15435875952243805, 0.020806124433875084, 0.23555229604244232, 0.06075655668973923, 0.07018083333969116, -0.03907359018921852, 0.17685799300670624, -0.1052674949169159, 0.011976814828813076, 0.2128676474094391, -0.03831172361969948, 0.06525631994009018, 0.11611197143793106, -0.01394710224121809, -0.0662488266825676, 0.036592915654182434, -0.009823341853916645, -0.07247381657361984, -0.21345274150371552, -0.1158827692270279, -0.1087421104311943, 0.054685093462467194, 0.04713849350810051, 0.050020426511764526, 0.1613347977399826, 0.07427749037742615, -0.04962149262428284, -0.0022197163198143244, 0.06106492131948471, 0.0832381621003151, 0.2504972517490387, -0.06253999471664429, 0.1427627056837082, -0.025090228766202927, -0.16789253056049347, 0.06259234994649887, 0.0661388710141182, 0.09291604906320572, 0.06118352338671684, 0.10224727541208267, 0.005179570056498051, 0.009344357997179031, 0.12825439870357513, 0.07115643471479416, 0.008030776865780354, -0.03595518320798874, -0.039997417479753494, -0.03642706945538521, -0.013250070624053478, 0.032193150371313095, 0.046790316700935364, -0.16567666828632355, -0.021018991246819496, 0.009807335212826729, 0.05824935808777809, 0.02185324765741825, 0.08615364134311676, -0.18498282134532928, -0.016169089823961258, 0.06576614826917648, -0.011832303367555141, -0.11644340306520462, 0.08480028808116913, 0.0007836486911401153, -0.1121063381433487, 0.03723234683275223, -0.027525627985596657, 0.13150714337825775, -0.08457524329423904, 0.0741792693734169, -0.12022519111633301, -0.0374552421271801, -0.010245736688375473, 0.12193918228149414, -0.29501426219940186, 0.19123348593711853, -0.009575535543262959, -0.04439779743552208, -0.1071409061551094, -0.015645509585738182, 0.02963484264910221, 0.10361164063215256, 0.11110331863164902, -0.020523378625512123, -0.02764100395143032, 0.06007368490099907, -0.07205203175544739, 0.0399978905916214, 0.09906689822673798, -0.06730470806360245, -0.013155711814761162, -0.052545808255672455, 0.00039069546619430184, 0.010376452468335629, -0.10966821759939194, 0.022783124819397926, -0.19194799661636353, 0.08703918755054474, 0.08162695169448853, 0.09630028903484344, 0.037212129682302475, -0.029887177050113678, -0.07769683748483658, 0.2589099109172821, 0.009560960344970226, -0.10013746470212936, -0.10953836888074875, 0.008171502500772476, 0.04785030707716942, -0.07699282467365265, -0.016966527327895164, -0.0694924145936966, 0.04450516775250435, -0.06552471220493317, -0.18611730635166168, 0.11722762882709503, -0.09691806137561798, -0.03250948712229729, -0.036249466240406036, 0.21333028376102448, -0.03155504912137985, 0.017869247123599052, 0.04537748545408249, -0.00578570831567049, -0.11741422116756439, -0.10654788464307785, 0.0012778750387951732, -0.004119161982089281, 0.016931969672441483, 0.023226622492074966, -0.03199922665953636, -0.009455137886106968, -0.06797713041305542, -0.014383019879460335, 0.3228513300418854, 0.12615877389907837, -0.042267147451639175, 0.15242800116539001, 0.09877358376979828, -0.06251336634159088, -0.2941497564315796, -0.11165541410446167, -0.07421603053808212, -0.05438753217458725, -0.09733224660158157, -0.18137554824352264, 0.08739634603261948, -0.05383281409740448, -0.013516134582459927, 0.09413999319076538, -0.25194358825683594, -0.10185287892818451, 0.2005643993616104, -0.03753361105918884, 0.4304826855659485, -0.11250142753124237, -0.07815388590097427, -0.04850279167294502, -0.14005880057811737, 0.19035954773426056, 0.004324326757341623, 0.10461755096912384, -0.0006430890643969178, 0.19764995574951172, 0.05591731518507004, -0.0006032987730577588, 0.07056128233671188, 0.01866593211889267, -0.057801030576229095, -0.09095179289579391, -0.0913778692483902, -0.0337459035217762, 0.010270410217344761, 0.0292131919413805, -0.07448325306177139, 0.04388400912284851, -0.13094636797904968, -0.05198022723197937, -0.08626694977283478, 0.038746368139982224, 0.027130719274282455, -0.06653520464897156, -0.0030553280375897884, -0.04914497584104538, 0.0004573945188894868, 0.007742773275822401, 0.21047258377075195, -0.10902713984251022, 0.1467881053686142, 0.028732312843203545, 0.1500566452741623, -0.09794784337282181, -0.04768699035048485, -0.06421241164207458, -0.05478411167860031, 0.07145597785711288, -0.12202182412147522, 0.03240978345274925, 0.1044924184679985, -0.026888413354754448, 0.08732181787490845, 0.1105954647064209, -0.010995322838425636, 0.005803761538118124, 0.08983830362558365, -0.241703063249588, -0.06713853776454926, -0.08410414308309555, 0.05373041704297066, 0.05893997475504875, 0.10275863856077194, 0.20927143096923828, 0.007167487405240536, -0.031165437772870064, 0.021489497274160385, 0.027375908568501472, -0.017840299755334854, 0.05977841466665268, 0.010519524104893208, 0.030491052195429802, -0.14741286635398865, 0.043485816568136215, -0.013757874257862568, -0.09077676385641098, 0.02600322663784027, 0.14754873514175415, -0.10901660472154617, -0.12182232737541199, -0.03921690955758095, 0.13600249588489532, -0.14775370061397552, -0.009947444312274456, -0.0477454848587513, -0.12692049145698547, 0.06857728958129883, 0.1067143976688385, 0.0457911379635334, 0.04121949151158333, -0.09239879250526428, -0.027268609032034874, -0.0535728819668293, 0.00003198942795279436, 0.028995376080274582, -0.0204177163541317, -0.05248761177062988, 0.040780652314424515, -0.03588524088263512, 0.12051229178905487, -0.08552545309066772, -0.10064204037189484, -0.16698434948921204, 0.03528384119272232, -0.07174701243638992, -0.08977310359477997, -0.0871967226266861, -0.03724304214119911, 0.006766482722014189, -0.0405125692486763, -0.02825779654085636, -0.03461418300867081, -0.1126255914568901, 0.03079685941338539, -0.04579872637987137, 0.003088617930188775, -0.07116411626338959, 0.029772473499178886, 0.0525958277285099, -0.029091687873005867, 0.149556964635849, 0.14025014638900757, -0.11192594468593597, 0.09547203034162521, -0.1507159322500229, -0.07066365331411362, 0.09605675935745239, 0.018403515219688416, 0.04981891065835953, 0.05175008252263069, 0.009065150283277035, 0.051755502820014954, 0.06169715151190758, 0.04307684674859047, 0.0153890922665596, -0.07590135186910629, 0.06697173416614532, -0.06090308725833893, -0.10307016223669052, -0.05066140368580818, -0.003966273739933968, 0.015159476548433304, 0.07283487915992737, 0.10097057372331619, -0.056661296635866165, 0.09506311267614365, -0.05649305135011673, 0.04625694453716278, 0.024318000301718712, -0.17797043919563293, 0.03397766128182411, -0.08718447387218475, 0.05030312016606331, 0.010050542652606964, 0.1727033108472824, 0.02054430916905403, -0.019508427008986473, 0.02473587542772293, 0.0719463899731636, 0.04261681064963341, -0.013226886279881, 0.19012948870658875, 0.10657399147748947, -0.03943915665149689, -0.0805516242980957, 0.09759991616010666, 0.04438556358218193, 0.04173632711172104, 0.14543114602565765, -0.05563090741634369, -0.03441290557384491, 0.081944540143013, -0.0026839920319616795, 0.010976077988743782, -0.09896437078714371, -0.13543705642223358, -0.026787811890244484, 0.036508288234472275, -0.03667739778757095, 0.10571453720331192, 0.15851758420467377, -0.005720720160752535, 0.01726081222295761, -0.01855739764869213, -0.05729815363883972, -0.1993623524904251, -0.19528920948505402, -0.083323635160923, -0.13647840917110443, 0.0050200955010950565, -0.13574683666229248, 0.04266147315502167, 0.026296362280845642, 0.09698255360126495, -0.04634363576769829, 0.050944969058036804, 0.03791060671210289, -0.11099781841039658, 0.058360110968351364, -0.043620482087135315, 0.09173028916120529, -0.03267880156636238, 0.014702340587973595, -0.060175783932209015, 0.035412851721048355, 0.016039982438087463, 0.041373249143362045, -0.02921622060239315, 0.019025372341275215, -0.12458328902721405, -0.08709227293729782, -0.06697598844766617, 0.06596853584051132, 0.006195025984197855, 0.16954803466796875, 0.019531596451997757, -0.027915386483073235, 0.028833186253905296, 0.23899038136005402, -0.07318265736103058, -0.09635625779628754, -0.06982157379388809, 0.21012257039546967, -0.009315763600170612, 0.08784335851669312, -0.03747710958123207, 0.009438461624085903, -0.08562079071998596, 0.3506644368171692, 0.29213622212409973, -0.09391074627637863, 0.010968702845275402, -0.0027621579356491566, 0.04181644320487976, 0.12788556516170502, 0.09239348024129868, 0.10824161767959595, 0.29070642590522766, -0.06708572804927826, -0.03647898510098457, -0.006994254421442747, -0.0254643727093935, -0.055716969072818756, 0.0551714263856411, 0.05315792188048363, -0.06511329114437103, -0.01592782698571682, 0.11738577485084534, -0.2489209920167923, 0.0614120177924633, -0.15840938687324524, -0.16190756857395172, -0.07126864790916443, -0.0001230158086400479, 0.0958227664232254, 0.01604771800339222, 0.09578458964824677, -0.011418631300330162, -0.06834693253040314, 0.04414822906255722, 0.020037546753883362, -0.20774760842323303, 0.009963343851268291, 0.06968449801206589, -0.051950447261333466, -0.05526239052414894, -0.017540784552693367, 0.07181108742952347, 0.0862373560667038, 0.031932324171066284, -0.021655123680830002, 0.04088883846998215, -0.011214682832360268, -0.07533704489469528, 0.03916772082448006, 0.027806051075458527, 0.005651058629155159, -0.08518505096435547, 0.07656224071979523, -0.16369622945785522, 0.03412613272666931, -0.0035786160733550787, -0.048953261226415634, -0.014727948233485222, 0.030175231397151947, -0.061420172452926636, 0.08509553223848343, 0.0839199498295784, -0.0171944722533226, -0.016525855287909508, -0.0222842525690794, -0.012990890070796013, -0.020874707028269768, -0.0818524956703186, -0.09698375314474106, -0.15574125945568085, -0.1261346936225891, 0.08575325459241867, -0.00355695397593081, -0.19997835159301758, 0.028783639892935753, -0.12125882506370544, 0.04249454662203789, -0.12142720073461533, 0.09701541811227798, 0.0825105607509613, 0.02303435280919075, -0.0030652873683720827, 0.006164520047605038, 0.03737448528409004, 0.07968182861804962, -0.13731823861598969, -0.08554888516664505 ]
null
null
null
# ResNet This ResNet18 model was translated from the ONNX ResNetv1 model found at https://github.com/onnx/models/tree/main/vision/classification/resnet into Axon using [AxonOnnx](https://github.com/elixir-nx/axon_onnx) The following description is copied from the relevant description at the ONNX repository. ## Use cases These ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required. ImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches. ## Description Deeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity. ## Model ResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers. ResNet v1 uses post-activation for the residual blocks. ### Input All pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (N x 3 x H x W), where N is the batch size, and H and W are expected to be at least 224. The inference was done using jpeg image. ### Preprocessing The images have to be loaded in to a range of [0, 1] and then normalized using mean = [0.485, 0.456, 0.406] and std = [0.229, 0.224, 0.225]. The transformation should preferably happen at preprocessing. ### Output The model outputs image scores for each of the 1000 classes of ImageNet. ### Postprocessing The post-processing involves calculating the softmax probability scores for each class. You can also sort them to report the most probable classes. Check [imagenet_postprocess.py](../imagenet_postprocess.py) for code. ## Dataset Dataset used for train and validation: [ImageNet (ILSVRC2012)](http://www.image-net.org/challenges/LSVRC/2012/). Check [imagenet_prep](../imagenet_prep.md) for guidelines on preparing the dataset. ## References * **ResNetv1** [Deep residual learning for image recognition](https://arxiv.org/abs/1512.03385) He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778. 2016. * **ONNX source model** [onnx/models vision/classification/resnet resnet18-v1-7.onnx](https://github.com/onnx/models/tree/main/vision/classification/resnet/README)
{"license": "apache-2.0", "tags": ["Axon", "Elixir"], "datasets": ["ImageNet"]}
null
Axon/resnet18-v1
[ "Axon", "Elixir", "dataset:ImageNet", "arxiv:1512.03385", "license:apache-2.0", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1512.03385" ]
[]
TAGS #Axon #Elixir #dataset-ImageNet #arxiv-1512.03385 #license-apache-2.0 #region-us
# ResNet This ResNet18 model was translated from the ONNX ResNetv1 model found at URL into Axon using AxonOnnx The following description is copied from the relevant description at the ONNX repository. ## Use cases These ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required. ImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches. ## Description Deeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity. ## Model ResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers. ResNet v1 uses post-activation for the residual blocks. ### Input All pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (N x 3 x H x W), where N is the batch size, and H and W are expected to be at least 224. The inference was done using jpeg image. ### Preprocessing The images have to be loaded in to a range of [0, 1] and then normalized using mean = [0.485, 0.456, 0.406] and std = [0.229, 0.224, 0.225]. The transformation should preferably happen at preprocessing. ### Output The model outputs image scores for each of the 1000 classes of ImageNet. ### Postprocessing The post-processing involves calculating the softmax probability scores for each class. You can also sort them to report the most probable classes. Check imagenet_postprocess.py for code. ## Dataset Dataset used for train and validation: ImageNet (ILSVRC2012). Check imagenet_prep for guidelines on preparing the dataset. ## References * ResNetv1 Deep residual learning for image recognition He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778. 2016. * ONNX source model onnx/models vision/classification/resnet URL
[ "# ResNet\r\n\r\nThis ResNet18 model was translated from the ONNX ResNetv1 model found \r\nat URL into Axon using AxonOnnx\r\nThe following description is copied from the relevant description at the ONNX repository.", "## Use cases\r\nThese ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required.\r\n\r\nImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches.", "## Description\r\nDeeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity.", "## Model\r\n\r\nResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers.\r\nResNet v1 uses post-activation for the residual blocks.", "### Input\r\nAll pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (N x 3 x H x W), where N is the batch size, and H and W are expected to be at least 224.\r\nThe inference was done using jpeg image.", "### Preprocessing\r\nThe images have to be loaded in to a range of [0, 1] and then normalized using mean = [0.485, 0.456, 0.406] and std = [0.229, 0.224, 0.225]. The transformation should preferably happen at preprocessing.", "### Output\r\nThe model outputs image scores for each of the 1000 classes of ImageNet.", "### Postprocessing\r\nThe post-processing involves calculating the softmax probability scores for each class. You can also sort them to report the most probable classes. Check imagenet_postprocess.py for code.", "## Dataset\r\nDataset used for train and validation: ImageNet (ILSVRC2012). Check imagenet_prep for guidelines on preparing the dataset.", "## References\r\n* ResNetv1\r\nDeep residual learning for image recognition\r\n He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778. 2016.\r\n\r\n* ONNX source model\r\nonnx/models vision/classification/resnet URL" ]
[ "TAGS\n#Axon #Elixir #dataset-ImageNet #arxiv-1512.03385 #license-apache-2.0 #region-us \n", "# ResNet\r\n\r\nThis ResNet18 model was translated from the ONNX ResNetv1 model found \r\nat URL into Axon using AxonOnnx\r\nThe following description is copied from the relevant description at the ONNX repository.", "## Use cases\r\nThese ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required.\r\n\r\nImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches.", "## Description\r\nDeeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity.", "## Model\r\n\r\nResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers.\r\nResNet v1 uses post-activation for the residual blocks.", "### Input\r\nAll pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (N x 3 x H x W), where N is the batch size, and H and W are expected to be at least 224.\r\nThe inference was done using jpeg image.", "### Preprocessing\r\nThe images have to be loaded in to a range of [0, 1] and then normalized using mean = [0.485, 0.456, 0.406] and std = [0.229, 0.224, 0.225]. The transformation should preferably happen at preprocessing.", "### Output\r\nThe model outputs image scores for each of the 1000 classes of ImageNet.", "### Postprocessing\r\nThe post-processing involves calculating the softmax probability scores for each class. You can also sort them to report the most probable classes. Check imagenet_postprocess.py for code.", "## Dataset\r\nDataset used for train and validation: ImageNet (ILSVRC2012). Check imagenet_prep for guidelines on preparing the dataset.", "## References\r\n* ResNetv1\r\nDeep residual learning for image recognition\r\n He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778. 2016.\r\n\r\n* ONNX source model\r\nonnx/models vision/classification/resnet URL" ]
[ 36, 51, 157, 140, 56, 75, 67, 21, 48, 35, 79 ]
[ "passage: TAGS\n#Axon #Elixir #dataset-ImageNet #arxiv-1512.03385 #license-apache-2.0 #region-us \n# ResNet\r\n\r\nThis ResNet18 model was translated from the ONNX ResNetv1 model found \r\nat URL into Axon using AxonOnnx\r\nThe following description is copied from the relevant description at the ONNX repository.## Use cases\r\nThese ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required.\r\n\r\nImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches.## Description\r\nDeeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity.## Model\r\n\r\nResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers.\r\nResNet v1 uses post-activation for the residual blocks." ]
[ -0.03244570642709732, 0.1143403872847557, -0.002907772082835436, 0.07804016023874283, 0.10426665097475052, 0.015642421320080757, 0.06919723749160767, 0.06106363236904144, -0.05125245451927185, 0.03717862069606781, 0.021020103245973587, -0.05612923949956894, 0.03542623668909073, 0.021515261381864548, 0.04324430599808693, -0.26741549372673035, 0.010686901398003101, -0.08474316447973251, 0.03801620751619339, 0.022571789100766182, 0.12170156091451645, -0.07437127828598022, 0.10465922951698303, 0.032630804926157, -0.0065165357664227486, 0.037587299942970276, -0.00012187870015623048, -0.020674053579568863, 0.1422262042760849, 0.06997888535261154, 0.13897587358951569, -0.0372563973069191, 0.14085249602794647, -0.19497469067573547, 0.04036925733089447, 0.09942980110645294, -0.004901749547570944, 0.07919245213270187, 0.0039009274914860725, 0.07779869437217712, 0.12980233132839203, 0.0019873129203915596, 0.05107017606496811, 0.04823358729481697, -0.0557430163025856, -0.13233913481235504, -0.08412633091211319, 0.11387304216623306, -0.003115191590040922, 0.056497782468795776, 0.018209248781204224, 0.09577851742506027, 0.022371551021933556, 0.016851263120770454, 0.10447154939174652, -0.2552465796470642, -0.0034019742161035538, 0.15369653701782227, -0.041410449892282486, 0.11488396674394608, -0.046254366636276245, 0.021913357079029083, 0.028726084157824516, 0.019016074016690254, 0.03761790692806244, 0.0383223220705986, 0.04145427793264389, -0.03196598216891289, -0.08499787002801895, -0.03762882202863693, 0.13950268924236298, 0.03884298726916313, -0.07225868105888367, -0.09387263655662537, -0.10849212110042572, 0.0545063242316246, 0.002389865927398205, -0.12063844501972198, 0.04157961905002594, 0.060370903462171555, 0.09544005244970322, -0.08083778619766235, -0.10296982526779175, 0.013022384606301785, -0.05803924426436424, 0.017985470592975616, 0.0648559108376503, 0.03803128004074097, 0.01477773580700159, 0.08624744415283203, -0.059308066964149475, 0.015243595466017723, -0.02717028371989727, -0.13289135694503784, -0.05831604450941086, 0.004686145577579737, -0.009462883695960045, -0.10487328469753265, -0.07423293590545654, 0.14283329248428345, -0.07406218349933624, 0.03067239746451378, -0.00007204454595921561, 0.028528137132525444, 0.0432281419634819, 0.060848355293273926, -0.06835032254457474, 0.007594357244670391, 0.03554104268550873, -0.011594715528190136, 0.0253151785582304, -0.051421426236629486, -0.06702610105276108, 0.0627717599272728, 0.0370013490319252, 0.010244810953736305, -0.05841058865189552, -0.0191390011459589, -0.00918720755726099, -0.052639324218034744, 0.18091297149658203, -0.03405407443642616, -0.037476249039173126, -0.04176926985383034, -0.06802907586097717, 0.10741104185581207, 0.08178902417421341, 0.020791377872228622, -0.04058705270290375, 0.0325341634452343, -0.0714632049202919, -0.0845140740275383, -0.03963514044880867, -0.09341917186975479, -0.00637378403916955, -0.006114120129495859, 0.003041696734726429, -0.2173117995262146, -0.17215462028980255, -0.026606349274516106, 0.0883033499121666, 0.02030259743332863, -0.006834207568317652, 0.05877650901675224, -0.025009725242853165, -0.0503447987139225, 0.0026024000253528357, -0.07005541771650314, -0.001951416488736868, -0.059297699481248856, -0.10890863090753555, 0.01856345124542713, -0.0035241360310465097, 0.008027425967156887, -0.022058486938476562, -0.0005303464131429791, -0.14393675327301025, 0.008325379341840744, -0.09563669562339783, -0.03819068521261215, -0.09690213203430176, -0.07886257022619247, -0.05762149766087532, 0.012029336765408516, 0.044275254011154175, 0.09911065548658371, -0.15105392038822174, 0.0033606458455324173, 0.029922138899564743, -0.13603858649730682, -0.010037618689239025, 0.1413581371307373, -0.08296135067939758, 0.06657731533050537, 0.09175462275743484, 0.06834413856267929, 0.047107432037591934, -0.07161904871463776, -0.12061517685651779, -0.04417192563414574, -0.08140522986650467, 0.05532758682966232, 0.0006566309602931142, 0.03468501195311546, 0.03949136659502983, -0.026048673316836357, 0.046773843467235565, -0.055566202849149704, -0.01760796830058098, -0.07330582290887833, -0.010836904868483543, -0.033777814358472824, 0.047548066824674606, -0.009566688910126686, 0.014455041848123074, 0.04734106734395027, -0.11981037259101868, -0.09808313101530075, 0.11168674379587173, -0.02210579439997673, 0.044526588171720505, -0.10731638967990875, 0.03712204098701477, -0.012503453530371189, 0.006080653052777052, -0.14077039062976837, -0.09363292902708054, 0.050493981689214706, -0.08764965087175369, 0.09704448282718658, 0.11232202500104904, 0.03144819661974907, 0.03704953193664551, -0.056316398084163666, 0.00291337794624269, -0.09037794172763824, -0.017253294587135315, -0.06653571873903275, -0.06445983797311783, -0.008350372314453125, -0.05427909269928932, 0.0024447059258818626, -0.21454021334648132, -0.004396345000714064, -0.06546708196401596, 0.07303199172019958, 0.07137646526098251, -0.03526555374264717, -0.01878846064209938, -0.058166563510894775, -0.04427073895931244, -0.06362957507371902, -0.012096296064555645, -0.020060613751411438, -0.05149321258068085, 0.0004907419788651168, -0.08952467143535614, 0.03022240847349167, 0.0276554636657238, 0.07911621034145355, -0.04675871506333351, -0.015361485071480274, -0.03707403317093849, 0.004045586567372084, -0.06196600943803787, 0.016221797093749046, 0.2742452621459961, 0.01747259311378002, 0.060587864369153976, -0.08465299755334854, 0.060835499316453934, 0.01779850944876671, -0.016884442418813705, 0.016287703067064285, -0.018009088933467865, 0.03424634784460068, -0.17795321345329285, -0.07134052366018295, 0.024177517741918564, 0.001125452108681202, 0.006016120780259371, 0.044816043227910995, -0.06709165871143341, -0.009870694950222969, -0.06659374386072159, 0.036530233919620514, 0.11162721365690231, 0.09061327576637268, 0.03387835994362831, 0.014662151224911213, 0.06109772250056267, 0.0423269085586071, -0.05058755725622177, 0.03550472855567932, 0.0253947414457798, 0.015866519883275032, -0.014111059717833996, -0.013686041347682476, -0.05411983281373978, 0.086614690721035, 0.013462900184094906, -0.03060484305024147, -0.07167018949985504, -0.034745972603559494, -0.0520600900053978, 0.1600407510995865, -0.08646218478679657, -0.14325949549674988, -0.12180374562740326, 0.047409795224666595, -0.03509491682052612, -0.003263725433498621, 0.011746455915272236, -0.12874284386634827, -0.125321626663208, -0.11213138699531555, 0.017935819923877716, -0.020292650908231735, -0.037977851927280426, 0.02283794805407524, 0.022093432024121284, 0.008086632937192917, -0.11618944257497787, 0.024026596918702126, -0.09985930472612381, -0.05897315591573715, -0.012107132002711296, -0.005554127972573042, 0.06513021886348724, 0.1013697162270546, -0.01039552129805088, -0.052800580859184265, 0.008378904312849045, 0.11168600618839264, -0.0465305857360363, 0.14347800612449646, 0.20564477145671844, -0.027023637667298317, 0.07648536562919617, 0.025635357946157455, 0.04265927895903587, -0.05455659702420235, 0.06899489462375641, 0.06150975078344345, -0.05988409370183945, -0.25745564699172974, -0.08972588926553726, 0.021176738664507866, -0.0555187463760376, 0.0539708286523819, 0.05435599759221077, 0.0022567890118807554, 0.06518301367759705, -0.048758409917354584, 0.024652257561683655, 0.016747191548347473, 0.05884476751089096, 0.11015838384628296, -0.016560962423682213, 0.07458312064409256, -0.05795605480670929, -0.015332911163568497, 0.12197147309780121, 0.08281944692134857, 0.27491310238838196, -0.011318924836814404, 0.02232486382126808, 0.04498772323131561, 0.03106852062046528, 0.04730920121073723, 0.08300910890102386, -0.003711606143042445, -0.004772163461893797, -0.08366264402866364, 0.021540870890021324, -0.020348867401480675, 0.11369737982749939, -0.009408973157405853, -0.05798419192433357, -0.024057887494564056, 0.1445978283882141, -0.006008084863424301, 0.07893387228250504, 0.010795291513204575, -0.15346243977546692, -0.1436760574579239, 0.019496576860547066, -0.07599899172782898, -0.07380329817533493, 0.06695926934480667, 0.10094591975212097, -0.10324659198522568, -0.023950256407260895, -0.06210847571492195, 0.07484270632266998, -0.15203891694545746, 0.01535346731543541, -0.07154521346092224, 0.08646558225154877, -0.015674063935875893, 0.06853850930929184, -0.04090331494808197, 0.12770001590251923, 0.01260446198284626, 0.07642773538827896, -0.05537494644522667, 0.0051244585774838924, 0.08874791115522385, 0.06533874571323395, 0.13567329943180084, 0.00752801401540637, -0.15242955088615417, 0.03499368950724602, -0.08074488490819931, 0.04091104865074158, 0.062083300203084946, -0.014387029223144054, 0.0692770928144455, -0.021866561844944954, 0.03739926964044571, 0.04061922803521156, 0.03817916288971901, -0.20869773626327515, -0.15240605175495148, 0.01979878731071949, 0.012838417664170265, -0.05920255556702614, -0.0688442662358284, -0.042978376150131226, -0.053508397191762924, 0.16118183732032776, -0.10060693323612213, -0.03965994715690613, -0.10245150327682495, -0.002410084242001176, 0.11598964780569077, -0.06350299715995789, 0.03722503036260605, 0.008324656635522842, 0.13568855822086334, -0.05072939023375511, -0.07467205822467804, 0.03004532866179943, -0.10524694621562958, -0.12966997921466827, -0.06100419536232948, -0.03166065737605095, 0.12984445691108704, 0.04973849281668663, 0.029889928176999092, 0.014701666310429573, -0.05823149159550667, -0.10627938061952591, 0.029466012492775917, 0.16457679867744446, -0.027499862015247345, 0.08199317008256912, -0.11832545697689056, 0.05568650737404823, -0.07555844634771347, 0.026075467467308044, 0.07297604531049728, 0.11748397350311279, -0.09595414996147156, 0.0704372376203537, 0.16603073477745056, -0.09473211318254471, -0.2093561440706253, -0.04890444502234459, -0.0195841696113348, 0.0029854588210582733, 0.06981703639030457, -0.21476803719997406, 0.10597143322229385, 0.013499469496309757, -0.01920556277036667, -0.09030734747648239, -0.13568232953548431, -0.09083132445812225, 0.09876146912574768, 0.0009198710322380066, 0.18963906168937683, -0.0005860136589035392, 0.016461502760648727, -0.04345684498548508, 0.10317886620759964, 0.06509879231452942, -0.10342033207416534, 0.06508708000183105, -0.015378771349787712, 0.06905286759138107, 0.010371532291173935, -0.01017868984490633, 0.07237064093351364, 0.04020312801003456, 0.0903412476181984, -0.03438837081193924, 0.06845664978027344, 0.03990420699119568, -0.07326571643352509, 0.1416400969028473, 0.13498501479625702, 0.09981514513492584, -0.11523962020874023, -0.04222749546170235, 0.014630874618887901, 0.06017842888832092, 0.00014860625378787518, -0.06943094730377197, -0.1338312327861786, 0.11468381434679031, 0.10389146953821182, 0.023722801357507706, 0.10720539838075638, 0.031844910234212875, 0.07820629328489304, 0.1108483299612999, 0.1164834126830101, -0.00169876660220325, -0.039323605597019196, 0.011866921558976173, -0.011187509633600712, 0.06239583343267441, -0.1867191046476364, 0.012398353777825832, 0.10920757800340652, -0.05699184909462929, 0.009511079639196396, 0.03431187942624092, -0.17095912992954254, 0.021993594244122505, 0.02652641013264656, -0.060783885419368744, -0.1661992073059082, -0.021991534158587456, 0.08528222888708115, -0.10103456676006317, 0.06202505901455879, 0.18520401418209076, -0.1269531548023224, 0.021730272099375725, 0.005241659469902515, 0.0646413266658783, 0.04827629029750824, 0.04479134827852249, 0.12264780700206757, 0.004617969039827585, -0.025720685720443726, 0.1864546686410904, 0.08880381286144257, -0.065545953810215, 0.03269360959529877, -0.018338095396757126, -0.06722190231084824, -0.05760061740875244, -0.10603409260511398, 0.03840111196041107, -0.12338797748088837, -0.08175934851169586, -0.04425870254635811, -0.014967283234000206, 0.005219286773353815, 0.004678846802562475, 0.028482109308242798, -0.026842443272471428, -0.10635178536176682, -0.002939108293503523, -0.09046373516321182, 0.05536554381251335, 0.03791891783475876, 0.03415398672223091, -0.1552342027425766, -0.06866134703159332, 0.0011003697291016579, 0.040613386780023575, -0.03342081606388092, -0.001930915517732501, -0.039258018136024475, -0.04392126202583313, -0.0805954560637474, -0.051303375512361526, 0.01035603228956461, 0.015389323234558105, -0.016761593520641327, 0.0058714342303574085, -0.03831948712468147, 0.05265183374285698, -0.024679621681571007, 0.003190360264852643, 0.0005176271079108119, 0.04618876799941063, -0.10585527122020721, 0.02917506918311119, 0.029401643201708794, -0.02960607223212719, 0.1286887526512146, -0.05646894499659538, -0.026667002588510513, 0.06099167466163635, -0.0672946497797966, 0.04857037216424942, -0.023167278617620468, 0.06879865378141403, 0.0037878884468227625, -0.08783599734306335, 0.012319103814661503, -0.052539389580488205, -0.033172838389873505, -0.03765736147761345, 0.015494994819164276, 0.0020699689630419016, -0.005888211540877819, -0.04775596037507057, 0.04918362572789192, -0.0645492747426033, -0.0019378955475986004, 0.04704345390200615, 0.05954140052199364, 0.10924134403467178, 0.026194356381893158, 0.017173076048493385, -0.14235946536064148, 0.010676704347133636, 0.03381233289837837, 0.006907213944941759, 0.05236120894551277, -0.059505805373191833, -0.0016646689036861062, -0.027931012213230133, 0.2337854504585266, -0.032709021121263504, -0.008104223757982254, 0.03560406342148781, 0.024854211136698723, -0.17576561868190765, -0.042190439999103546, -0.09205576777458191, 0.07263623923063278, 0.003845165017992258, -0.011897816322743893, -0.04589315876364708, 0.013897507451474667, -0.016853362321853638, 0.15733946859836578, 0.07637453824281693, 0.06976263225078583, 0.07105692476034164, -0.0243026465177536, -0.07889454811811447, -0.048092879354953766, 0.07397075742483139, -0.05276932567358017, 0.01246108952909708, -0.024809425696730614, -0.09957017004489899, 0.15874408185482025, -0.0360267348587513, 0.08542885631322861, -0.009436514228582382, -0.04875899851322174, -0.11698811501264572, -0.06482008099555969, -0.06496747583150864, -0.09549859166145325, 0.0006176149472594261, -0.1197967380285263, -0.02069198526442051, 0.113101065158844, -0.009814533405005932, -0.0350847952067852, 0.07069764286279678, -0.11155115067958832, -0.060650888830423355, 0.0703592523932457, -0.0024243216030299664, -0.03784840181469917, 0.10460304468870163, -0.011129072867333889, 0.0018008614424616098, 0.02641231380403042, 0.036909591406583786, 0.029650002717971802, 0.08574944734573364, 0.05339507758617401, 0.012615914456546307, -0.03665049746632576, -0.02963036671280861, -0.008596225641667843, -0.035868726670742035, 0.1505810022354126, 0.01546009536832571, -0.032794296741485596, -0.007411458063870668, 0.22064392268657684, -0.043580178171396255, 0.06781059503555298, -0.1549520641565323, 0.19558273255825043, 0.03292674571275711, 0.03132665157318115, -0.023273233324289322, -0.07309374958276749, 0.013268277049064636, 0.11468446254730225, 0.11865939944982529, -0.10527654737234116, -0.03633204847574234, 0.035783059895038605, -0.010789348743855953, -0.053717441856861115, 0.11067875474691391, 0.06749799847602844, 0.19822187721729279, -0.030736129730939865, 0.05356518551707268, -0.026278600096702576, 0.036338578909635544, -0.04885520413517952, 0.10276617854833603, 0.006700250320136547, 0.02097339741885662, -0.06118374317884445, 0.10636412352323532, -0.09399108588695526, -0.27765384316444397, -0.03714807704091072, -0.08726615458726883, -0.07791397720575333, 0.005393812898546457, 0.0031835727859288454, -0.014821589924395084, 0.09096161276102066, -0.007807521615177393, -0.009141352027654648, 0.1694989949464798, 0.016801325604319572, -0.023348288610577583, -0.010226249694824219, 0.03573082014918327, 0.04728863388299942, 0.20317624509334564, 0.030612360686063766, 0.012874373234808445, 0.07676460593938828, -0.06117095425724983, -0.09840940684080124, 0.027073519304394722, -0.001013689674437046, -0.03124777227640152, -0.0020708756055682898, 0.11883861571550369, -0.005043590907007456, 0.10384956002235413, 0.07558562606573105, -0.06771094352006912, 0.0373663604259491, 0.07037226110696793, -0.08105397969484329, -0.03678077459335327, 0.09743650257587433, -0.11470472067594528, 0.1723310947418213, 0.07954133301973343, -0.014294310472905636, -0.0018016285030171275, -0.025666404515504837, 0.023471418768167496, 0.008833409287035465, 0.08668879419565201, 0.0466667115688324, -0.061662815511226654, -0.0062750536017119884, -0.13196592032909393, 0.061073943972587585, -0.1834908425807953, -0.05127450078725815, 0.0330188162624836, -0.02413448877632618, -0.02170444466173649, 0.07292815297842026, 0.08295516669750214, -0.016412975266575813, -0.025753110647201538, 0.07985714823007584, -0.009885837323963642, 0.02111150324344635, -0.10842403024435043, -0.05262899026274681 ]
null
null
null
# ResNet This ResNet34 model was translated from the ONNX ResNetv1 model found at https://github.com/onnx/models/tree/main/vision/classification/resnet into Axon using [AxonOnnx](https://github.com/elixir-nx/axon_onnx) The following description is copied from the relevant description at the ONNX repository. ## Use cases These ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required. ImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches. ## Description Deeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity. ## Model ResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers. ResNet v1 uses post-activation for the residual blocks. ### Input All pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (N x 3 x H x W), where N is the batch size, and H and W are expected to be at least 224. The inference was done using jpeg image. ### Preprocessing The images have to be loaded in to a range of [0, 1] and then normalized using mean = [0.485, 0.456, 0.406] and std = [0.229, 0.224, 0.225]. The transformation should preferably happen at preprocessing. ### Output The model outputs image scores for each of the 1000 classes of ImageNet. ### Postprocessing The post-processing involves calculating the softmax probability scores for each class. You can also sort them to report the most probable classes. Check [imagenet_postprocess.py](../imagenet_postprocess.py) for code. ## Dataset Dataset used for train and validation: [ImageNet (ILSVRC2012)](http://www.image-net.org/challenges/LSVRC/2012/). Check [imagenet_prep](../imagenet_prep.md) for guidelines on preparing the dataset. ## References * **ResNetv1** [Deep residual learning for image recognition](https://arxiv.org/abs/1512.03385) He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778. 2016. * **ONNX source model** [onnx/models vision/classification/resnet resnet34-v1-7.onnx](https://github.com/onnx/models/tree/main/vision/classification/resnet/README)
{"license": "apache-2.0", "tags": ["Axon", "Elixir"], "datasets": ["ImageNet"]}
null
Axon/resnet34-v1
[ "Axon", "Elixir", "dataset:ImageNet", "arxiv:1512.03385", "license:apache-2.0", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1512.03385" ]
[]
TAGS #Axon #Elixir #dataset-ImageNet #arxiv-1512.03385 #license-apache-2.0 #region-us
# ResNet This ResNet34 model was translated from the ONNX ResNetv1 model found at URL into Axon using AxonOnnx The following description is copied from the relevant description at the ONNX repository. ## Use cases These ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required. ImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches. ## Description Deeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity. ## Model ResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers. ResNet v1 uses post-activation for the residual blocks. ### Input All pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (N x 3 x H x W), where N is the batch size, and H and W are expected to be at least 224. The inference was done using jpeg image. ### Preprocessing The images have to be loaded in to a range of [0, 1] and then normalized using mean = [0.485, 0.456, 0.406] and std = [0.229, 0.224, 0.225]. The transformation should preferably happen at preprocessing. ### Output The model outputs image scores for each of the 1000 classes of ImageNet. ### Postprocessing The post-processing involves calculating the softmax probability scores for each class. You can also sort them to report the most probable classes. Check imagenet_postprocess.py for code. ## Dataset Dataset used for train and validation: ImageNet (ILSVRC2012). Check imagenet_prep for guidelines on preparing the dataset. ## References * ResNetv1 Deep residual learning for image recognition He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778. 2016. * ONNX source model onnx/models vision/classification/resnet URL
[ "# ResNet\r\n\r\nThis ResNet34 model was translated from the ONNX ResNetv1 model found \r\nat URL into Axon using AxonOnnx\r\nThe following description is copied from the relevant description at the ONNX repository.", "## Use cases\r\nThese ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required.\r\n\r\nImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches.", "## Description\r\nDeeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity.", "## Model\r\n\r\nResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers.\r\nResNet v1 uses post-activation for the residual blocks.", "### Input\r\nAll pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (N x 3 x H x W), where N is the batch size, and H and W are expected to be at least 224.\r\nThe inference was done using jpeg image.", "### Preprocessing\r\nThe images have to be loaded in to a range of [0, 1] and then normalized using mean = [0.485, 0.456, 0.406] and std = [0.229, 0.224, 0.225]. The transformation should preferably happen at preprocessing.", "### Output\r\nThe model outputs image scores for each of the 1000 classes of ImageNet.", "### Postprocessing\r\nThe post-processing involves calculating the softmax probability scores for each class. You can also sort them to report the most probable classes. Check imagenet_postprocess.py for code.", "## Dataset\r\nDataset used for train and validation: ImageNet (ILSVRC2012). Check imagenet_prep for guidelines on preparing the dataset.", "## References\r\n* ResNetv1\r\nDeep residual learning for image recognition\r\n He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778. 2016.\r\n\r\n* ONNX source model\r\nonnx/models vision/classification/resnet URL" ]
[ "TAGS\n#Axon #Elixir #dataset-ImageNet #arxiv-1512.03385 #license-apache-2.0 #region-us \n", "# ResNet\r\n\r\nThis ResNet34 model was translated from the ONNX ResNetv1 model found \r\nat URL into Axon using AxonOnnx\r\nThe following description is copied from the relevant description at the ONNX repository.", "## Use cases\r\nThese ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required.\r\n\r\nImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches.", "## Description\r\nDeeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity.", "## Model\r\n\r\nResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers.\r\nResNet v1 uses post-activation for the residual blocks.", "### Input\r\nAll pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (N x 3 x H x W), where N is the batch size, and H and W are expected to be at least 224.\r\nThe inference was done using jpeg image.", "### Preprocessing\r\nThe images have to be loaded in to a range of [0, 1] and then normalized using mean = [0.485, 0.456, 0.406] and std = [0.229, 0.224, 0.225]. The transformation should preferably happen at preprocessing.", "### Output\r\nThe model outputs image scores for each of the 1000 classes of ImageNet.", "### Postprocessing\r\nThe post-processing involves calculating the softmax probability scores for each class. You can also sort them to report the most probable classes. Check imagenet_postprocess.py for code.", "## Dataset\r\nDataset used for train and validation: ImageNet (ILSVRC2012). Check imagenet_prep for guidelines on preparing the dataset.", "## References\r\n* ResNetv1\r\nDeep residual learning for image recognition\r\n He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778. 2016.\r\n\r\n* ONNX source model\r\nonnx/models vision/classification/resnet URL" ]
[ 36, 51, 157, 140, 56, 75, 67, 21, 48, 35, 79 ]
[ "passage: TAGS\n#Axon #Elixir #dataset-ImageNet #arxiv-1512.03385 #license-apache-2.0 #region-us \n# ResNet\r\n\r\nThis ResNet34 model was translated from the ONNX ResNetv1 model found \r\nat URL into Axon using AxonOnnx\r\nThe following description is copied from the relevant description at the ONNX repository.## Use cases\r\nThese ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required.\r\n\r\nImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches.## Description\r\nDeeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity.## Model\r\n\r\nResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers.\r\nResNet v1 uses post-activation for the residual blocks." ]
[ -0.03160631284117699, 0.11817073822021484, -0.002887010807171464, 0.0792945846915245, 0.10384716838598251, 0.016046935692429543, 0.06844524294137955, 0.061475664377212524, -0.05602551996707916, 0.03668763488531113, 0.0216473750770092, -0.05524459108710289, 0.03333454579114914, 0.023008067160844803, 0.04387406259775162, -0.26924511790275574, 0.008696865290403366, -0.08448749035596848, 0.036604806780815125, 0.022517060860991478, 0.12120118737220764, -0.07463569194078445, 0.10497058928012848, 0.03181500732898712, -0.004558233078569174, 0.038342103362083435, -0.0012815421214327216, -0.019894514232873917, 0.1415017992258072, 0.06824363023042679, 0.13837659358978271, -0.03839122876524925, 0.14245478808879852, -0.19553643465042114, 0.039991024881601334, 0.09772154688835144, -0.004228348843753338, 0.07882223278284073, 0.0024777317885309458, 0.07765696197748184, 0.12731069326400757, 0.0004978944198228419, 0.0501343309879303, 0.047223981469869614, -0.056150294840335846, -0.13511499762535095, -0.0849594697356224, 0.11395320296287537, -0.006331311538815498, 0.05680447071790695, 0.019035354256629944, 0.09799958765506744, 0.021311568096280098, 0.01754639483988285, 0.10250944644212723, -0.25692135095596313, -0.0017895711353048682, 0.1561463624238968, -0.04205533117055893, 0.11231810599565506, -0.049378134310245514, 0.022791270166635513, 0.028297068551182747, 0.018314942717552185, 0.03719834238290787, 0.03644604608416557, 0.042457934468984604, -0.030698232352733612, -0.0852241963148117, -0.036703526973724365, 0.1406029462814331, 0.03880372270941734, -0.07251640409231186, -0.09276828169822693, -0.1078108623623848, 0.05458955466747284, 0.002857151674106717, -0.1186099573969841, 0.04271680489182472, 0.060162320733070374, 0.09159328788518906, -0.08060136437416077, -0.10249153524637222, 0.012861782684922218, -0.05891665816307068, 0.01760883629322052, 0.06457384675741196, 0.03819579258561134, 0.01633165217936039, 0.08838062733411789, -0.0600743293762207, 0.015162096358835697, -0.027915509417653084, -0.13250653445720673, -0.05573992803692818, 0.003607163205742836, -0.010532923974096775, -0.09933531284332275, -0.07356168329715729, 0.14672718942165375, -0.07290346175432205, 0.03030325472354889, 0.0037825091276317835, 0.02681203931570053, 0.043535757809877396, 0.060403697192668915, -0.06657693535089493, 0.0033119164872914553, 0.03744016960263252, -0.010831651277840137, 0.026703115552663803, -0.04993550479412079, -0.06742583215236664, 0.062424637377262115, 0.034878432750701904, 0.01070414762943983, -0.05934106186032295, -0.01634928397834301, -0.009122820571064949, -0.051681049168109894, 0.17888431251049042, -0.034675177186727524, -0.035992950201034546, -0.0405627004802227, -0.0693395659327507, 0.10501115769147873, 0.08297599852085114, 0.020362507551908493, -0.04051036387681961, 0.0292076263576746, -0.07146112620830536, -0.08416147530078888, -0.04033184424042702, -0.09386037290096283, -0.006263410672545433, -0.0065038008615374565, 0.0028146947734057903, -0.2176101952791214, -0.17331279814243317, -0.02592172473669052, 0.08755194395780563, 0.021081898361444473, -0.006905569229274988, 0.05705840513110161, -0.025435389950871468, -0.050187848508358, 0.0025184203404933214, -0.06917792558670044, -0.0027567551005631685, -0.05859356373548508, -0.10596003383398056, 0.019242340698838234, -0.005896258633583784, 0.007178943604230881, -0.022652819752693176, -0.0008616026025265455, -0.1428622007369995, 0.007765001151710749, -0.09437330812215805, -0.03869390860199928, -0.09642572700977325, -0.08045732975006104, -0.05670206993818283, 0.010888532735407352, 0.04381336271762848, 0.09993098676204681, -0.14670637249946594, 0.004814548883587122, 0.028206951916217804, -0.13408789038658142, -0.007190048228949308, 0.13965347409248352, -0.08423410356044769, 0.0691651850938797, 0.09281334280967712, 0.06756988912820816, 0.04760027304291725, -0.07195116579532623, -0.11944681406021118, -0.04250358045101166, -0.08237604051828384, 0.05598260089755058, 0.0009901553858071566, 0.03445954620838165, 0.0378061980009079, -0.025289112702012062, 0.04679694026708603, -0.05513659864664078, -0.017500165849924088, -0.07369983941316605, -0.010143427178263664, -0.03436460718512535, 0.048605985939502716, -0.010401594452559948, 0.015655122697353363, 0.048309847712516785, -0.11883585900068283, -0.09597183018922806, 0.11143925786018372, -0.023001328110694885, 0.04446413740515709, -0.10636386275291443, 0.03753930330276489, -0.012440977618098259, 0.006013411562889814, -0.14066526293754578, -0.09574797749519348, 0.05066021904349327, -0.08703732490539551, 0.0978359580039978, 0.11304272711277008, 0.032169047743082047, 0.03913282975554466, -0.05747637525200844, 0.004315152298659086, -0.08986762911081314, -0.01737980730831623, -0.06595016270875931, -0.06456483155488968, -0.009233839809894562, -0.054699406027793884, 0.0031581916846334934, -0.21452262997627258, -0.004748012870550156, -0.06486181914806366, 0.07399264723062515, 0.07077211886644363, -0.03592817857861519, -0.017290623858571053, -0.05712876841425896, -0.04333508387207985, -0.06374167650938034, -0.010146444663405418, -0.020058540627360344, -0.0523635633289814, -0.0003965698415413499, -0.09001006186008453, 0.03149896487593651, 0.02816757932305336, 0.0785614624619484, -0.04799215495586395, -0.017215535044670105, -0.03693527728319168, 0.004337357357144356, -0.06051405891776085, 0.015420985408127308, 0.274421751499176, 0.014697051607072353, 0.06189190223813057, -0.08470028638839722, 0.058843910694122314, 0.015323513187468052, -0.017250878736376762, 0.017676763236522675, -0.017494361847639084, 0.032642047852277756, -0.18052496016025543, -0.07062418758869171, 0.021432029083371162, -0.0007139000226743519, 0.006847730837762356, 0.044539131224155426, -0.06673532724380493, -0.011455066502094269, -0.06562657654285431, 0.03586233779788017, 0.11331731826066971, 0.09019835293292999, 0.034007176756858826, 0.015070524998009205, 0.06109555438160896, 0.04140346869826317, -0.05187411233782768, 0.03579669073224068, 0.024483900517225266, 0.015181602910161018, -0.019077882170677185, -0.01342805102467537, -0.05393408611416817, 0.08562266081571579, 0.014683213084936142, -0.029719922691583633, -0.07051169127225876, -0.03422962874174118, -0.05155296251177788, 0.15985427796840668, -0.08760400116443634, -0.1433393955230713, -0.12344648689031601, 0.044433366507291794, -0.03496624156832695, -0.00305720092728734, 0.011266405694186687, -0.12777179479599, -0.1263306885957718, -0.11206619441509247, 0.019931409507989883, -0.01947605051100254, -0.03768349811434746, 0.022444674745202065, 0.023478658869862556, 0.008168112486600876, -0.11723323911428452, 0.024522600695490837, -0.10020069777965546, -0.05840599909424782, -0.012312345206737518, -0.0058864932507276535, 0.0652410089969635, 0.09903336316347122, -0.00968613289296627, -0.05258909985423088, 0.008292351849377155, 0.1142221987247467, -0.04565136134624481, 0.14285726845264435, 0.2042800635099411, -0.027086082845926285, 0.0768468976020813, 0.02541634626686573, 0.04337934032082558, -0.05523643270134926, 0.06971802562475204, 0.06255734711885452, -0.05885336175560951, -0.2542187571525574, -0.08793359994888306, 0.020969728007912636, -0.056823182851076126, 0.05365023389458656, 0.05450020730495453, 0.0002979465643875301, 0.06438715010881424, -0.047440171241760254, 0.02579590119421482, 0.017556674778461456, 0.06008285656571388, 0.11404068768024445, -0.018046077340841293, 0.07421588152647018, -0.05888202786445618, -0.018077721819281578, 0.12197906523942947, 0.08389104902744293, 0.2739114761352539, -0.01224552746862173, 0.02273857407271862, 0.04387165606021881, 0.03410498425364494, 0.046744897961616516, 0.08320912718772888, -0.004037891514599323, -0.0032292029354721308, -0.08331887423992157, 0.020902518182992935, -0.019135186448693275, 0.11298973858356476, -0.005830138456076384, -0.05795234814286232, -0.023262372240424156, 0.1406574696302414, -0.006456498987972736, 0.07687367498874664, 0.012290595099329948, -0.15669016540050507, -0.1439773142337799, 0.018972087651491165, -0.07531854510307312, -0.07244009524583817, 0.06595931202173233, 0.10078638046979904, -0.10381115972995758, -0.022420769557356834, -0.060926005244255066, 0.07411939650774002, -0.15242023766040802, 0.01474459283053875, -0.07023199647665024, 0.08727128058671951, -0.015070631168782711, 0.06829921901226044, -0.0366598516702652, 0.12610091269016266, 0.012983006425201893, 0.07634524255990982, -0.0546722337603569, 0.00458537507802248, 0.08993194997310638, 0.06645072251558304, 0.13385556638240814, 0.00743643706664443, -0.14624172449111938, 0.030693622305989265, -0.0822814479470253, 0.04068445414304733, 0.06518438458442688, -0.015084787271916866, 0.07212108373641968, -0.022624975070357323, 0.03660738840699196, 0.04060230776667595, 0.034364551305770874, -0.2090706080198288, -0.15186557173728943, 0.01921885646879673, 0.013077172450721264, -0.05748942494392395, -0.06777670234441757, -0.043809037655591965, -0.05419573560357094, 0.16113442182540894, -0.0996248871088028, -0.039004672318696976, -0.10197429358959198, -0.00015678901399951428, 0.11775685101747513, -0.0627356693148613, 0.03751010447740555, 0.007111868821084499, 0.13209141790866852, -0.04828878864645958, -0.07473189383745193, 0.028961826115846634, -0.10299394279718399, -0.13041441142559052, -0.06051937863230705, -0.030891111120581627, 0.12832821905612946, 0.04903443530201912, 0.030631007626652718, 0.015509207732975483, -0.058931123465299606, -0.10702387988567352, 0.027865884825587273, 0.16339245438575745, -0.02340245060622692, 0.08314058184623718, -0.12190678715705872, 0.05883364751935005, -0.0747467651963234, 0.02758573740720749, 0.07482393831014633, 0.11660774797201157, -0.09533839672803879, 0.07124655693769455, 0.16360507905483246, -0.09514590352773666, -0.21175557374954224, -0.04902714490890503, -0.019795062020421028, 0.0033873103093355894, 0.06819534301757812, -0.21544866263866425, 0.10650565475225449, 0.011104290373623371, -0.01828751713037491, -0.08826636523008347, -0.1350536048412323, -0.09124426543712616, 0.09584183990955353, 0.0016160408267751336, 0.18927153944969177, -0.0009772934718057513, 0.015844350680708885, -0.04322432726621628, 0.10257349163293839, 0.0646570473909378, -0.1032753586769104, 0.06857890635728836, -0.01532998587936163, 0.06995853036642075, 0.010146738030016422, -0.010540934279561043, 0.07381214946508408, 0.039511702954769135, 0.09049870073795319, -0.035764146596193314, 0.06724274158477783, 0.043568532913923264, -0.07306960970163345, 0.14246152341365814, 0.1359378695487976, 0.10069406032562256, -0.11494181305170059, -0.04337943345308304, 0.01460140012204647, 0.06073026359081268, 0.0004607397713698447, -0.06877310574054718, -0.13369104266166687, 0.11485368013381958, 0.10428542643785477, 0.022894233465194702, 0.10624893009662628, 0.032361164689064026, 0.07446707040071487, 0.1099390834569931, 0.11631978303194046, 0.0024451324716210365, -0.04101712629199028, 0.01333011593669653, -0.01189684309065342, 0.06149149313569069, -0.1841706782579422, 0.01110681239515543, 0.10922016203403473, -0.05512991547584534, 0.008239967748522758, 0.034747228026390076, -0.16899146139621735, 0.02250501699745655, 0.027827640995383263, -0.06151023879647255, -0.16520550847053528, -0.0210017878562212, 0.0779879242181778, -0.10124747455120087, 0.06334862858057022, 0.1864139288663864, -0.1254773885011673, 0.022266710177063942, 0.005572186317294836, 0.0633767694234848, 0.04648292437195778, 0.044910598546266556, 0.123085618019104, 0.0038088245783001184, -0.02587232179939747, 0.18715430796146393, 0.08818793296813965, -0.06361416727304459, 0.03188162297010422, -0.01841835491359234, -0.06924951821565628, -0.05713643506169319, -0.10577680915594101, 0.04120495170354843, -0.122219018638134, -0.08106637746095657, -0.046485453844070435, -0.013660281896591187, 0.00514261657372117, 0.006526353303343058, 0.029591664671897888, -0.02553137205541134, -0.10803855955600739, -0.0015362570993602276, -0.0902809202671051, 0.05522293969988823, 0.03900454565882683, 0.03310674801468849, -0.15469974279403687, -0.06572753190994263, 0.0010275347158312798, 0.03994603082537651, -0.03409223258495331, -0.0031490421388298273, -0.04083499312400818, -0.04494287818670273, -0.07696353644132614, -0.05136577785015106, 0.009880133904516697, 0.01684555970132351, -0.017672831192612648, 0.005975297652184963, -0.03780543804168701, 0.05411593243479729, -0.02440854161977768, 0.0017116462113335729, 0.0006946602370589972, 0.045539770275354385, -0.10625126957893372, 0.028182117268443108, 0.029871072620153427, -0.029354846104979515, 0.12988756597042084, -0.05504161864519119, -0.026948414742946625, 0.06106157600879669, -0.06562716513872147, 0.04874764755368233, -0.02174912393093109, 0.0691804513335228, 0.0020374658051878214, -0.08983743190765381, 0.012281462550163269, -0.053209125995635986, -0.030776571482419968, -0.036660365760326385, 0.021626396104693413, 0.001589729217812419, -0.007756910752505064, -0.04843731224536896, 0.0492263026535511, -0.0646887794137001, -0.0027252123691141605, 0.04619212821125984, 0.058884553611278534, 0.11006734520196915, 0.0262310728430748, 0.015447716228663921, -0.14176738262176514, 0.009993551298975945, 0.03402609005570412, 0.005624276120215654, 0.04730142652988434, -0.05917288362979889, -0.0006289971643127501, -0.026285095140337944, 0.23486541211605072, -0.034027647227048874, -0.011813781224191189, 0.03425874933600426, 0.025919780135154724, -0.17252306640148163, -0.04303855821490288, -0.08849631994962692, 0.07291156053543091, 0.0023460914380848408, -0.011588389985263348, -0.04340822994709015, 0.014692428521811962, -0.019890399649739265, 0.15591232478618622, 0.07804755121469498, 0.07336364686489105, 0.07119669020175934, -0.02399560995399952, -0.07920579612255096, -0.04748072475194931, 0.07479880750179291, -0.050670865923166275, 0.013139782473444939, -0.02701525017619133, -0.09936903417110443, 0.15494252741336823, -0.035542555153369904, 0.08439798653125763, -0.009005053900182247, -0.050022002309560776, -0.11666862666606903, -0.06557610630989075, -0.06493867933750153, -0.09466362744569778, 0.0009341734694316983, -0.12000983208417892, -0.02049056999385357, 0.11115339398384094, -0.010119299404323101, -0.035012491047382355, 0.07153132557868958, -0.11752983182668686, -0.06154235452413559, 0.06895861029624939, -0.0027934610843658447, -0.03835257515311241, 0.10668622702360153, -0.01228928193449974, 0.0011299487669020891, 0.0248484555631876, 0.036370061337947845, 0.02952236495912075, 0.08676360547542572, 0.05295995995402336, 0.012193256989121437, -0.035283658653497696, -0.030413251370191574, -0.006432886701077223, -0.034167610108852386, 0.15247885882854462, 0.01541446428745985, -0.03344954177737236, -0.00793639849871397, 0.2205432951450348, -0.042774002999067307, 0.06583999842405319, -0.15587040781974792, 0.19622215628623962, 0.032731398940086365, 0.03151069954037666, -0.02265692502260208, -0.0732576847076416, 0.01405797153711319, 0.11596261709928513, 0.12106142193078995, -0.10655184835195541, -0.036581579595804214, 0.03692685812711716, -0.01064965222030878, -0.05085330083966255, 0.11022759974002838, 0.06763055175542831, 0.19755776226520538, -0.030656473711133003, 0.04988263547420502, -0.02509777992963791, 0.03719809278845787, -0.049028363078832626, 0.10166679322719574, 0.006292316131293774, 0.01993359439074993, -0.06346863508224487, 0.10524490475654602, -0.09481589496135712, -0.2770649790763855, -0.034170594066381454, -0.08827074617147446, -0.07771804928779602, 0.005615298170596361, 0.00178526621311903, -0.016489291563630104, 0.09012430161237717, -0.007275096140801907, -0.008101935498416424, 0.17261697351932526, 0.016729092225432396, -0.02543690986931324, -0.010618740692734718, 0.03627263009548187, 0.04815975949168205, 0.20091865956783295, 0.02995491400361061, 0.016675999388098717, 0.0780336856842041, -0.061839550733566284, -0.09810145199298859, 0.027996456250548363, -0.0008506164303980768, -0.03361884132027626, -0.0030256742611527443, 0.11707107722759247, -0.005400299094617367, 0.10610165446996689, 0.07242835313081741, -0.06464347243309021, 0.038271673023700714, 0.07404784858226776, -0.0814092606306076, -0.03844323754310608, 0.09534693509340286, -0.11644644290208817, 0.17295274138450623, 0.08093326538801193, -0.013428943231701851, -0.001323421485722065, -0.02483206056058407, 0.024517247453331947, 0.009460587054491043, 0.07941846549510956, 0.04566659405827522, -0.06253521889448166, -0.007039657793939114, -0.13369952142238617, 0.06040535494685173, -0.18308335542678833, -0.05359905958175659, 0.03222798928618431, -0.022652680054306984, -0.020430414006114006, 0.07360061258077621, 0.08136450499296188, -0.01616821251809597, -0.025462649762630463, 0.07501200586557388, -0.009578412398695946, 0.020520826801657677, -0.11101671308279037, -0.05453097075223923 ]
null
null
null
# ResNet This ResNet50 model was translated from the ONNX ResNetv1 model found at https://github.com/onnx/models/tree/main/vision/classification/resnet into Axon using [AxonOnnx](https://github.com/elixir-nx/axon_onnx) The following description is copied from the relevant description at the ONNX repository. ## Use cases These ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required. ImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches. ## Description Deeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity. ## Model ResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers. ResNet v1 uses post-activation for the residual blocks. ### Input All pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (N x 3 x H x W), where N is the batch size, and H and W are expected to be at least 224. The inference was done using jpeg image. ### Preprocessing The images have to be loaded in to a range of [0, 1] and then normalized using mean = [0.485, 0.456, 0.406] and std = [0.229, 0.224, 0.225]. The transformation should preferably happen at preprocessing. ### Output The model outputs image scores for each of the 1000 classes of ImageNet. ### Postprocessing The post-processing involves calculating the softmax probability scores for each class. You can also sort them to report the most probable classes. Check [imagenet_postprocess.py](../imagenet_postprocess.py) for code. ## Dataset Dataset used for train and validation: [ImageNet (ILSVRC2012)](http://www.image-net.org/challenges/LSVRC/2012/). Check [imagenet_prep](../imagenet_prep.md) for guidelines on preparing the dataset. ## References * **ResNetv1** [Deep residual learning for image recognition](https://arxiv.org/abs/1512.03385) He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778. 2016. * **ONNX source model** [onnx/models vision/classification/resnet resnet50-v1-7.onnx](https://github.com/onnx/models/tree/main/vision/classification/resnet/README)
{"license": "apache-2.0", "tags": ["Axon", "Elixir"], "datasets": ["ImageNet"]}
null
Axon/resnet50-v1
[ "Axon", "Elixir", "dataset:ImageNet", "arxiv:1512.03385", "license:apache-2.0", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1512.03385" ]
[]
TAGS #Axon #Elixir #dataset-ImageNet #arxiv-1512.03385 #license-apache-2.0 #region-us
# ResNet This ResNet50 model was translated from the ONNX ResNetv1 model found at URL into Axon using AxonOnnx The following description is copied from the relevant description at the ONNX repository. ## Use cases These ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required. ImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches. ## Description Deeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity. ## Model ResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers. ResNet v1 uses post-activation for the residual blocks. ### Input All pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (N x 3 x H x W), where N is the batch size, and H and W are expected to be at least 224. The inference was done using jpeg image. ### Preprocessing The images have to be loaded in to a range of [0, 1] and then normalized using mean = [0.485, 0.456, 0.406] and std = [0.229, 0.224, 0.225]. The transformation should preferably happen at preprocessing. ### Output The model outputs image scores for each of the 1000 classes of ImageNet. ### Postprocessing The post-processing involves calculating the softmax probability scores for each class. You can also sort them to report the most probable classes. Check imagenet_postprocess.py for code. ## Dataset Dataset used for train and validation: ImageNet (ILSVRC2012). Check imagenet_prep for guidelines on preparing the dataset. ## References * ResNetv1 Deep residual learning for image recognition He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778. 2016. * ONNX source model onnx/models vision/classification/resnet URL
[ "# ResNet\r\n\r\nThis ResNet50 model was translated from the ONNX ResNetv1 model found \r\nat URL into Axon using AxonOnnx\r\nThe following description is copied from the relevant description at the ONNX repository.", "## Use cases\r\nThese ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required.\r\n\r\nImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches.", "## Description\r\nDeeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity.", "## Model\r\n\r\nResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers.\r\nResNet v1 uses post-activation for the residual blocks.", "### Input\r\nAll pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (N x 3 x H x W), where N is the batch size, and H and W are expected to be at least 224.\r\nThe inference was done using jpeg image.", "### Preprocessing\r\nThe images have to be loaded in to a range of [0, 1] and then normalized using mean = [0.485, 0.456, 0.406] and std = [0.229, 0.224, 0.225]. The transformation should preferably happen at preprocessing.", "### Output\r\nThe model outputs image scores for each of the 1000 classes of ImageNet.", "### Postprocessing\r\nThe post-processing involves calculating the softmax probability scores for each class. You can also sort them to report the most probable classes. Check imagenet_postprocess.py for code.", "## Dataset\r\nDataset used for train and validation: ImageNet (ILSVRC2012). Check imagenet_prep for guidelines on preparing the dataset.", "## References\r\n* ResNetv1\r\nDeep residual learning for image recognition\r\n He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778. 2016.\r\n\r\n* ONNX source model\r\nonnx/models vision/classification/resnet URL" ]
[ "TAGS\n#Axon #Elixir #dataset-ImageNet #arxiv-1512.03385 #license-apache-2.0 #region-us \n", "# ResNet\r\n\r\nThis ResNet50 model was translated from the ONNX ResNetv1 model found \r\nat URL into Axon using AxonOnnx\r\nThe following description is copied from the relevant description at the ONNX repository.", "## Use cases\r\nThese ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required.\r\n\r\nImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches.", "## Description\r\nDeeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity.", "## Model\r\n\r\nResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers.\r\nResNet v1 uses post-activation for the residual blocks.", "### Input\r\nAll pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (N x 3 x H x W), where N is the batch size, and H and W are expected to be at least 224.\r\nThe inference was done using jpeg image.", "### Preprocessing\r\nThe images have to be loaded in to a range of [0, 1] and then normalized using mean = [0.485, 0.456, 0.406] and std = [0.229, 0.224, 0.225]. The transformation should preferably happen at preprocessing.", "### Output\r\nThe model outputs image scores for each of the 1000 classes of ImageNet.", "### Postprocessing\r\nThe post-processing involves calculating the softmax probability scores for each class. You can also sort them to report the most probable classes. Check imagenet_postprocess.py for code.", "## Dataset\r\nDataset used for train and validation: ImageNet (ILSVRC2012). Check imagenet_prep for guidelines on preparing the dataset.", "## References\r\n* ResNetv1\r\nDeep residual learning for image recognition\r\n He, Kaiming, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778. 2016.\r\n\r\n* ONNX source model\r\nonnx/models vision/classification/resnet URL" ]
[ 36, 51, 157, 140, 56, 75, 67, 21, 48, 35, 79 ]
[ "passage: TAGS\n#Axon #Elixir #dataset-ImageNet #arxiv-1512.03385 #license-apache-2.0 #region-us \n# ResNet\r\n\r\nThis ResNet50 model was translated from the ONNX ResNetv1 model found \r\nat URL into Axon using AxonOnnx\r\nThe following description is copied from the relevant description at the ONNX repository.## Use cases\r\nThese ResNet models perform image classification - they take images as input and classify the major object in the image into a set of pre-defined classes. They are trained on ImageNet dataset which contains images from 1000 classes. ResNet models provide very high accuracies with affordable model sizes. They are ideal for cases when high accuracy of classification is required.\r\n\r\nImageNet trained models are often used as the base layers for a transfer learning approach to training a model in your domain. Transfer learning can significantly reduce the processing necessary to train an accurate model in your domain. This model was published here with the expectation that it would be useful to the Elixir community for transfer learning and other similar approaches.## Description\r\nDeeper neural networks are more difficult to train. Residual learning framework ease the training of networks that are substantially deeper. The research explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. It also provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset the residual nets were evaluated with a depth of up to 152 layers — 8× deeper than VGG nets but still having lower complexity.## Model\r\n\r\nResNet models consists of residual blocks and came up to counter the effect of deteriorating accuracies with more layers due to network not learning the initial layers.\r\nResNet v1 uses post-activation for the residual blocks." ]
[ -0.03223172947764397, 0.11707670241594315, -0.002874913392588496, 0.07761047035455704, 0.10437411069869995, 0.0165263582020998, 0.06847279518842697, 0.06109214574098587, -0.05489890277385712, 0.03651267662644386, 0.0207914300262928, -0.055109746754169464, 0.03597894310951233, 0.02345067635178566, 0.042580872774124146, -0.2686353623867035, 0.009400486014783382, -0.08392269909381866, 0.03424876555800438, 0.022541169077157974, 0.12176914513111115, -0.07610499113798141, 0.1053626537322998, 0.03251471370458603, -0.006185149773955345, 0.03739045187830925, -0.0014152942458167672, -0.0200803279876709, 0.1404716968536377, 0.06880417466163635, 0.13894592225551605, -0.03718681260943413, 0.14216828346252441, -0.19633732736110687, 0.040496163070201874, 0.09840086102485657, -0.003616263624280691, 0.07954349368810654, 0.004430519882589579, 0.07822997123003006, 0.12509451806545258, 0.0016983154928311706, 0.04973416402935982, 0.047520339488983154, -0.05503538250923157, -0.13013489544391632, -0.08427421748638153, 0.11346697807312012, -0.005095167551189661, 0.05612883344292641, 0.01932867243885994, 0.09652280062437057, 0.02044486068189144, 0.01667178049683571, 0.1055024117231369, -0.25725358724594116, -0.0017055210191756487, 0.15735137462615967, -0.0428638830780983, 0.11230538040399551, -0.04892025887966156, 0.02242070436477661, 0.027559887617826462, 0.017147080972790718, 0.03686770051717758, 0.03769857808947563, 0.04261765256524086, -0.03126448392868042, -0.08572147786617279, -0.03602493181824684, 0.14169089496135712, 0.038912028074264526, -0.07195492833852768, -0.09299129992723465, -0.10799071937799454, 0.053030893206596375, 0.0023593627847731113, -0.12138151377439499, 0.04294835776090622, 0.06023717299103737, 0.09263882040977478, -0.07997988164424896, -0.10238877683877945, 0.012468473054468632, -0.059580519795417786, 0.014611280523240566, 0.06479893624782562, 0.03815968707203865, 0.013280258513987064, 0.08857826143503189, -0.06300212442874908, 0.016154231503605843, -0.02933521382510662, -0.1334545761346817, -0.05678260698914528, 0.0023606009781360626, -0.010471915826201439, -0.10611878335475922, -0.07335664331912994, 0.14534442126750946, -0.07121619582176208, 0.030208628624677658, 0.00317202415317297, 0.027789399027824402, 0.0448765866458416, 0.060303427278995514, -0.0678112804889679, 0.0014275548746809363, 0.036694835871458054, -0.011261730454862118, 0.024302110075950623, -0.049928516149520874, -0.06717430055141449, 0.06148940697312355, 0.03652384504675865, 0.010151856578886509, -0.05930480360984802, -0.01728084683418274, -0.00908817257732153, -0.05251261591911316, 0.1780095249414444, -0.033681340515613556, -0.03666262701153755, -0.041054822504520416, -0.06884557753801346, 0.10724327713251114, 0.08080143481492996, 0.02207006886601448, -0.040742527693510056, 0.03196953609585762, -0.07087157666683197, -0.08318693935871124, -0.04108617082238197, -0.09434466809034348, -0.007477298378944397, -0.007318251300603151, 0.002760590286925435, -0.21695536375045776, -0.17251572012901306, -0.02636655420064926, 0.08824162185192108, 0.021193988621234894, -0.006797885522246361, 0.05649296194314957, -0.02458949387073517, -0.05018816143274307, 0.003289866726845503, -0.06758175045251846, -0.0019153500907123089, -0.059616848826408386, -0.10739294439554214, 0.019119834527373314, -0.003952311351895332, 0.007626559119671583, -0.021976834163069725, -0.0016194203635677695, -0.14260917901992798, 0.008781238459050655, -0.0932486429810524, -0.03963235393166542, -0.09721915423870087, -0.07884986698627472, -0.05974501371383667, 0.009766128845512867, 0.04379615560173988, 0.1001402884721756, -0.15056560933589935, 0.004907488357275724, 0.03012893535196781, -0.13421058654785156, -0.008178952150046825, 0.13893510401248932, -0.0836213231086731, 0.06883145868778229, 0.09182674437761307, 0.06694016605615616, 0.04679180681705475, -0.07285841554403305, -0.11946931481361389, -0.04326874390244484, -0.08238675445318222, 0.05580144375562668, 0.001042643329128623, 0.033476345241069794, 0.038750506937503815, -0.0256018228828907, 0.044844839721918106, -0.05611516162753105, -0.0177193321287632, -0.0741395428776741, -0.011184575036168098, -0.03535286337137222, 0.04651119187474251, -0.011309911496937275, 0.016786590218544006, 0.05006817355751991, -0.11717512458562851, -0.09787143766880035, 0.11056043207645416, -0.022320745512843132, 0.045151062309741974, -0.10649272799491882, 0.03665824979543686, -0.013463061302900314, 0.006632926408201456, -0.1422635316848755, -0.09213229268789291, 0.051409993320703506, -0.0905318483710289, 0.09728090465068817, 0.11288455873727798, 0.03238501399755478, 0.03823167085647583, -0.05687887966632843, 0.004272275138646364, -0.08984004706144333, -0.01688106171786785, -0.06665420532226562, -0.06367964297533035, -0.008064823225140572, -0.05427062138915062, 0.001150553347542882, -0.21293261647224426, -0.004361759405583143, -0.06534454226493835, 0.07423874735832214, 0.07235348969697952, -0.035692714154720306, -0.01791851408779621, -0.05772930756211281, -0.044744495302438736, -0.06345472484827042, -0.011013692244887352, -0.02086985856294632, -0.052337855100631714, -0.0026389427948743105, -0.08846312761306763, 0.03176412731409073, 0.02726886421442032, 0.07943399250507355, -0.047494880855083466, -0.019678551703691483, -0.036274973303079605, 0.004585080314427614, -0.06285813450813293, 0.017012707889080048, 0.27225103974342346, 0.016144681721925735, 0.060486625880002975, -0.08401746302843094, 0.06017319858074188, 0.015807490795850754, -0.015941355377435684, 0.017029503360390663, -0.019162815064191818, 0.03597088158130646, -0.1807512640953064, -0.07171878218650818, 0.02424975112080574, -0.0001800745667424053, 0.0054680160246789455, 0.0448208823800087, -0.06775795668363571, -0.011322233825922012, -0.06663741171360016, 0.03591626137495041, 0.11282234638929367, 0.09021314978599548, 0.03409648686647415, 0.015504350885748863, 0.06151249632239342, 0.041078392416238785, -0.05149218440055847, 0.03632344678044319, 0.02487056329846382, 0.015810629352927208, -0.019265661016106606, -0.013086296617984772, -0.05387053266167641, 0.08533868938684464, 0.014165523461997509, -0.028202632442116737, -0.07085946202278137, -0.034833211451768875, -0.05092248693108559, 0.1596820205450058, -0.0874294638633728, -0.14401288330554962, -0.12007676810026169, 0.0476631224155426, -0.034525126218795776, -0.003684982191771269, 0.011103757657110691, -0.12786731123924255, -0.12644918262958527, -0.11205209791660309, 0.020136410370469093, -0.0197959803044796, -0.03589651733636856, 0.02166888862848282, 0.02429911307990551, 0.006262172479182482, -0.11674147844314575, 0.025455376133322716, -0.10084164887666702, -0.05847826972603798, -0.011776327155530453, -0.005097066517919302, 0.06547992676496506, 0.10004998743534088, -0.00982683151960373, -0.050986338406801224, 0.008148439228534698, 0.11462409049272537, -0.046132560819387436, 0.1426466405391693, 0.20242619514465332, -0.025688299909234047, 0.07721301168203354, 0.027710294350981712, 0.0432293638586998, -0.05417972803115845, 0.06935083121061325, 0.06126578897237778, -0.05756809189915657, -0.25381577014923096, -0.08631942421197891, 0.02080433815717697, -0.05713721737265587, 0.05356362462043762, 0.053765710443258286, 0.001497361110523343, 0.06568510830402374, -0.04785111919045448, 0.024231310933828354, 0.018126213923096657, 0.05957341939210892, 0.10933999717235565, -0.016206910833716393, 0.07407918572425842, -0.057796914130449295, -0.017192471772432327, 0.12089263647794724, 0.08482881635427475, 0.27553436160087585, -0.010508266277611256, 0.022023923695087433, 0.04315458983182907, 0.034461576491594315, 0.04844227805733681, 0.08396108448505402, -0.0036482501309365034, -0.004205512348562479, -0.08319057524204254, 0.021817145869135857, -0.01866733282804489, 0.11299541592597961, -0.004149300511926413, -0.0577242374420166, -0.022326402366161346, 0.14335359632968903, -0.006070488132536411, 0.07851452380418777, 0.010847854427993298, -0.15268497169017792, -0.14259664714336395, 0.01862058788537979, -0.07506959140300751, -0.07288578897714615, 0.06624611467123032, 0.09925366938114166, -0.10458609461784363, -0.0233650840818882, -0.06269349902868271, 0.07480411976575851, -0.15165890753269196, 0.014826526865363121, -0.07209038734436035, 0.08817590028047562, -0.015579658560454845, 0.06793678551912308, -0.04329940676689148, 0.12480528652667999, 0.013149700127542019, 0.07693002372980118, -0.05536592751741409, 0.0042136781848967075, 0.08880115300416946, 0.06778661906719208, 0.13411056995391846, 0.006750169675797224, -0.14918677508831024, 0.033161576837301254, -0.0818595141172409, 0.039854343980550766, 0.06599852442741394, -0.015864429995417595, 0.0712711364030838, -0.022342966869473457, 0.03663180023431778, 0.04058980569243431, 0.03666507452726364, -0.21239390969276428, -0.15378743410110474, 0.020909132435917854, 0.0124538978561759, -0.05559912323951721, -0.06791701167821884, -0.0432661697268486, -0.054005205631256104, 0.15883037447929382, -0.10680623352527618, -0.03788385167717934, -0.10221637785434723, 0.00022244107094593346, 0.11597483605146408, -0.0628499910235405, 0.03807478025555611, 0.007244823034852743, 0.1342080682516098, -0.049136169254779816, -0.07442978024482727, 0.029665546491742134, -0.10470148921012878, -0.1293298453092575, -0.06167181208729744, -0.03310113772749901, 0.12854470312595367, 0.049734506756067276, 0.030496468767523766, 0.015161016024649143, -0.059288982301950455, -0.10570292174816132, 0.027447886765003204, 0.1662459373474121, -0.0253930501639843, 0.08401688933372498, -0.1223410815000534, 0.05588661506772041, -0.07545378804206848, 0.025875534862279892, 0.07527266442775726, 0.11567003279924393, -0.09538743644952774, 0.07225880026817322, 0.16320671141147614, -0.09452864527702332, -0.2109573483467102, -0.047598689794540405, -0.01802712492644787, 0.003783673280850053, 0.069539375603199, -0.21519462764263153, 0.10673205554485321, 0.01074332743883133, -0.01820754073560238, -0.08887979388237, -0.13752301037311554, -0.09167303144931793, 0.09769678115844727, 0.0018203879008069634, 0.19003579020500183, -0.0007586487336084247, 0.016392791643738747, -0.0436408706009388, 0.10220915824174881, 0.0678107738494873, -0.10301657021045685, 0.06769115477800369, -0.015300876460969448, 0.06888002902269363, 0.010369905270636082, -0.011136572808027267, 0.07201503217220306, 0.040364447981119156, 0.09084545075893402, -0.03484615311026573, 0.06690818071365356, 0.04323820397257805, -0.07351811230182648, 0.1413184553384781, 0.13663461804389954, 0.10022084414958954, -0.11054909229278564, -0.04287569969892502, 0.0158490352332592, 0.06144263595342636, 0.0008535631932318211, -0.06943385303020477, -0.13416822254657745, 0.11464548110961914, 0.1049344539642334, 0.023427259176969528, 0.10861929506063461, 0.03279658034443855, 0.07338462024927139, 0.10946514457464218, 0.11713044345378876, 0.0012325949501246214, -0.04180983453989029, 0.012445764616131783, -0.011611218564212322, 0.06230822578072548, -0.18171237409114838, 0.011559948325157166, 0.108864925801754, -0.05542493239045143, 0.008395560085773468, 0.0349867083132267, -0.1686623990535736, 0.021512040868401527, 0.026713499799370766, -0.060206130146980286, -0.1676216423511505, -0.020350908860564232, 0.08249499648809433, -0.1019175723195076, 0.06376737356185913, 0.1848694533109665, -0.1256382167339325, 0.022164111956954002, 0.0055924453772604465, 0.06442484259605408, 0.04848400130867958, 0.04641929641366005, 0.12274324893951416, 0.003759265411645174, -0.02595614828169346, 0.1882101446390152, 0.0884135365486145, -0.06520836055278778, 0.03160969913005829, -0.021086251363158226, -0.06997000426054001, -0.05656050890684128, -0.10511703044176102, 0.041368983685970306, -0.1233748123049736, -0.0810433030128479, -0.04498738795518875, -0.013494587503373623, 0.0046805390156805515, 0.004426116589456797, 0.0291118361055851, -0.02630634419620037, -0.10778475552797318, -0.0024077955167740583, -0.08980736136436462, 0.056062668561935425, 0.03870150446891785, 0.0342102013528347, -0.154964879155159, -0.06723602861166, 0.0008709580288268626, 0.03987240791320801, -0.033820852637290955, -0.0024898741394281387, -0.03958320617675781, -0.04547189548611641, -0.07491039484739304, -0.05115170404314995, 0.011003633961081505, 0.016234327107667923, -0.016891933977603912, 0.005311209242790937, -0.03786780685186386, 0.053474560379981995, -0.024670470505952835, 0.0021959198638796806, -0.000535077415406704, 0.04479273781180382, -0.10499496012926102, 0.02894890308380127, 0.030281182378530502, -0.02902635559439659, 0.12846806645393372, -0.05521271377801895, -0.027471452951431274, 0.060988813638687134, -0.06835409253835678, 0.04795345664024353, -0.02342851832509041, 0.06962516158819199, 0.0025160429067909718, -0.08899889886379242, 0.011866513639688492, -0.052515409886837006, -0.030910473316907883, -0.0373840406537056, 0.020646927878260612, 0.0015736849745735526, -0.00872254092246294, -0.05025705695152283, 0.05077045410871506, -0.06481466442346573, -0.0026284968480467796, 0.04632202163338661, 0.05922317132353783, 0.11196297407150269, 0.027163084596395493, 0.015758123248815536, -0.14028339087963104, 0.010777703486382961, 0.03452901914715767, 0.006525443866848946, 0.04724378138780594, -0.05932806804776192, -0.0005503824213519692, -0.026661759242415428, 0.2359374761581421, -0.034378767013549805, -0.010085874237120152, 0.034548569470644, 0.025499822571873665, -0.17380966246128082, -0.04335620254278183, -0.08694370836019516, 0.07547592371702194, 0.0027300859801471233, -0.01181117631494999, -0.042824290692806244, 0.014200064353644848, -0.01804688200354576, 0.15870243310928345, 0.077349953353405, 0.06817129999399185, 0.07151883840560913, -0.02626737393438816, -0.07993502169847488, -0.04456814005970955, 0.0772412121295929, -0.05124352127313614, 0.012481926009058952, -0.026344358921051025, -0.09529189020395279, 0.15667590498924255, -0.03555060550570488, 0.08412298560142517, -0.008457634598016739, -0.04974236711859703, -0.1179436445236206, -0.06666436046361923, -0.06562826782464981, -0.09552077949047089, 0.0007785194902680814, -0.11967910081148148, -0.019396813586354256, 0.11034911125898361, -0.009500742889940739, -0.03449525684118271, 0.07142183929681778, -0.11599858105182648, -0.0618138313293457, 0.06965196132659912, -0.0028575933538377285, -0.039298150688409805, 0.105862095952034, -0.012602428905665874, 0.0015396794769912958, 0.023851070553064346, 0.036761146038770676, 0.02862056903541088, 0.08581143617630005, 0.05192779004573822, 0.013336284086108208, -0.03576844558119774, -0.03006262145936489, -0.0077992030419409275, -0.03596055507659912, 0.15095557272434235, 0.01541647408157587, -0.03252670541405678, -0.007964597083628178, 0.2214461863040924, -0.04292961582541466, 0.06588060408830643, -0.15550242364406586, 0.19099459052085876, 0.03544173389673233, 0.03292127326130867, -0.022143831476569176, -0.07465176284313202, 0.013711666688323021, 0.11568474769592285, 0.11964588612318039, -0.10671310871839523, -0.03669751062989235, 0.037205129861831665, -0.01076984591782093, -0.05175968259572983, 0.1080261766910553, 0.06733052432537079, 0.19629575312137604, -0.03081820160150528, 0.053592998534440994, -0.02423691377043724, 0.036549583077430725, -0.04912801831960678, 0.1018667221069336, 0.006843808107078075, 0.019933270290493965, -0.06306855380535126, 0.1056690588593483, -0.09399320185184479, -0.2777252495288849, -0.037188321352005005, -0.08674251288175583, -0.07768025249242783, 0.004592245444655418, 0.0027747845742851496, -0.016162030398845673, 0.08936784416437149, -0.007022105623036623, -0.008208813145756721, 0.17296497523784637, 0.017346758395433426, -0.024740450084209442, -0.01153548900038004, 0.03564779832959175, 0.04610124230384827, 0.1994057148694992, 0.030910981819033623, 0.01523035578429699, 0.07719064503908157, -0.06149408221244812, -0.09782916307449341, 0.027237631380558014, -0.0011154162930324674, -0.03133717551827431, -0.0019446208607405424, 0.11928404867649078, -0.005369097460061312, 0.10528801381587982, 0.07301079481840134, -0.06651393324136734, 0.038529157638549805, 0.07142873853445053, -0.0809633657336235, -0.03850409388542175, 0.09784850478172302, -0.11450294405221939, 0.17290036380290985, 0.07955065369606018, -0.013363066129386425, -0.000515999214258045, -0.025994397699832916, 0.024756377562880516, 0.00870286300778389, 0.082693912088871, 0.045724719762802124, -0.06201210245490074, -0.006518873851746321, -0.13410384953022003, 0.060885827988386154, -0.18012017011642456, -0.053344085812568665, 0.03158660978078842, -0.0234291423112154, -0.020534571260213852, 0.07260145246982574, 0.0846472755074501, -0.016234220936894417, -0.02604657970368862, 0.07600204646587372, -0.009370370768010616, 0.02053729258477688, -0.11064337193965912, -0.053876928985118866 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # albert_bert_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "albert_bert_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/albert_bert_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# albert_bert_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
[ "# albert_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# albert_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ 62, 39, 6, 12, 8, 3, 118, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# albert_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ -0.08509203791618347, 0.1271563470363617, -0.003627235535532236, 0.06995399296283722, 0.1418028324842453, 0.038870152086019516, 0.1112576499581337, 0.1061035692691803, -0.11217713356018066, 0.0541842095553875, 0.0767379030585289, 0.07117237895727158, 0.04078949615359306, 0.15895403921604156, -0.03647996857762337, -0.21138975024223328, -0.003993574529886246, -0.004001093562692404, -0.06901755183935165, 0.11727892607450485, 0.08292553573846817, -0.10711560398340225, 0.05869891121983528, -0.005877855699509382, -0.13981708884239197, 0.019577737897634506, -0.03561665490269661, -0.0527474470436573, 0.09515223652124405, -0.005033911671489477, 0.09744206815958023, 0.037459589540958405, 0.14897887408733368, -0.19727562367916107, 0.004665984306484461, 0.10008805245161057, 0.04367423430085182, 0.09451564401388168, 0.05822642147541046, -0.01271052286028862, 0.12150890380144119, -0.15712618827819824, 0.10606309026479721, 0.017164772376418114, -0.08126740902662277, -0.10709492117166519, -0.10135296732187271, 0.053587645292282104, 0.0885876789689064, 0.12339286506175995, 0.005147561896592379, 0.1327328085899353, -0.11240323632955551, 0.0809164047241211, 0.18452179431915283, -0.20554867386817932, -0.04715605452656746, 0.05687316134572029, 0.021311188116669655, 0.06603023409843445, -0.09053989499807358, -0.015993431210517883, 0.029214780777692795, 0.023117540404200554, 0.10226500034332275, 0.004726466257125139, -0.09724145382642746, -0.0008403509273193777, -0.12733952701091766, -0.018927941098809242, 0.058921780437231064, 0.024147886782884598, -0.012230881489813328, -0.09100761264562607, -0.054842494428157806, -0.15212658047676086, -0.02685045637190342, -0.009682349860668182, 0.03811388835310936, -0.05965004116296768, -0.07238110154867172, -0.03844624012708664, -0.06296635419130325, -0.04776598885655403, -0.03124554455280304, 0.1312202513217926, 0.04334237426519394, 0.00797568541020155, -0.04656143859028816, 0.1150478944182396, 0.07717961817979813, -0.12400270998477936, 0.024277297779917717, 0.013901585713028908, -0.0898919627070427, -0.04775237292051315, -0.03232654184103012, -0.007912833243608475, -0.013178853318095207, 0.12844204902648926, -0.07700030505657196, 0.07693804800510406, 0.02624046802520752, -0.012526934035122395, -0.015419626608490944, 0.14641934633255005, -0.05075414106249809, -0.032935503870248795, -0.011736026033759117, 0.11587049812078476, 0.004508549347519875, -0.017331227660179138, -0.06523466855287552, 0.010195862501859665, 0.06575024127960205, 0.057952117174863815, -0.044204551726579666, 0.04483656957745552, -0.020657509565353394, -0.016210848465561867, 0.04228220880031586, -0.13635991513729095, 0.04473337158560753, 0.017485707998275757, -0.10084745287895203, -0.004871437791734934, 0.00722639262676239, -0.0012451369548216462, -0.04791664332151413, 0.1414465308189392, -0.07487507909536362, 0.007690786849707365, -0.08397834748029709, -0.08099470287561417, 0.009608154185116291, -0.11238637566566467, -0.03730636090040207, -0.0369967557489872, -0.1619693636894226, -0.06260241568088531, 0.024759948253631592, -0.05282921344041824, -0.03990894556045532, -0.05668968707323074, -0.07369979470968246, 0.01997734233736992, -0.013335194438695908, 0.18531647324562073, -0.07011403143405914, 0.06727756559848785, -0.033439986407756805, 0.02903478592634201, 0.059228070080280304, 0.045687828212976456, -0.06284778565168381, -0.014400424435734749, -0.09451916068792343, 0.08513644337654114, -0.07995297014713287, 0.003931113053113222, -0.10294681787490845, -0.09304970502853394, 0.0324409082531929, -0.005910682491958141, 0.0628366470336914, 0.12563735246658325, -0.20253820717334747, -0.03615693375468254, 0.10410071909427643, -0.04484190419316292, -0.00445045530796051, 0.052780766040086746, -0.05870063975453377, -0.014376374892890453, 0.06501462310552597, 0.12153372168540955, 0.05713193863630295, -0.12717019021511078, -0.002873749239370227, 0.0007992186001501977, 0.03874291107058525, 0.012589187361299992, 0.014243497513234615, 0.010768766514956951, 0.05442749336361885, 0.005591971799731255, -0.035982292145490646, 0.008568706922233105, -0.10447826236486435, -0.06374480575323105, -0.031575776636600494, -0.07735072076320648, 0.015481517650187016, 0.026834074407815933, 0.031956396996974945, -0.061428263783454895, -0.11484283208847046, 0.07483752071857452, 0.11890239268541336, -0.05065828189253807, 0.019317826256155968, -0.07636763900518417, -0.03655648231506348, 0.016408471390604973, -0.022192731499671936, -0.20145076513290405, -0.1047048270702362, 0.02080957591533661, -0.05499820411205292, 0.043035153299570084, -0.010845571756362915, 0.05908265337347984, 0.04248237609863281, -0.03590292111039162, -0.016264714300632477, -0.06480132788419724, 0.0028687804006040096, -0.09618488699197769, -0.21511591970920563, -0.030242884531617165, -0.015623281709849834, 0.16015346348285675, -0.21145080029964447, 0.010868254117667675, -0.028323277831077576, 0.1411549150943756, 0.024505745619535446, -0.05658017098903656, 0.005523674190044403, 0.06186213716864586, 0.0111498748883605, -0.09099218249320984, 0.03729625418782234, -0.004180997610092163, -0.07462369650602341, -0.01491566002368927, -0.14189137518405914, -0.021072180941700935, 0.08650337904691696, 0.07515563815832138, -0.1068037748336792, 0.02978692390024662, -0.07446452975273132, -0.03570731356739998, -0.09455147385597229, 0.04625747725367546, 0.2239859402179718, 0.03451031073927879, 0.13378052413463593, -0.04706117510795593, -0.08360453695058823, -0.007071619387716055, 0.021946556866168976, 0.018127700313925743, 0.11001677066087723, 0.08485547453165054, -0.03456132486462593, 0.06807361543178558, 0.02205396443605423, -0.043310243636369705, 0.1208912581205368, -0.04167856276035309, -0.08491861075162888, 0.007887258194386959, -0.038841333240270615, -0.026524584740400314, 0.09504877030849457, -0.028955737128853798, 0.007494454272091389, 0.029507329687476158, 0.02745942585170269, 0.03213445842266083, -0.19237276911735535, 0.00012156811862951145, 0.017432784661650658, -0.05005167797207832, -0.050476476550102234, -0.021784506738185883, 0.05867175757884979, 0.10923491418361664, 0.0070801652036607265, -0.032734937965869904, 0.0029103392735123634, -0.006981701124459505, -0.07020802050828934, 0.17198893427848816, -0.11275976151227951, -0.16030822694301605, -0.06972888857126236, 0.00670609762892127, -0.027107564732432365, -0.05205150693655014, 0.007099703419953585, -0.12194801867008209, -0.07007875293493271, -0.11597523838281631, -0.02007567137479782, 0.014506894163787365, -7.916304980426503e-7, 0.05667964369058609, 0.010742822661995888, 0.04642333835363388, -0.1360958367586136, 0.01280237827450037, -0.06036952883005142, -0.07686903327703476, 0.021041689440608025, 0.08688218891620636, 0.05925161764025688, 0.1526603400707245, -0.02069690078496933, 0.026970813050866127, -0.01865842193365097, 0.20680314302444458, -0.08928947150707245, 0.006027633789926767, 0.09972070902585983, 0.01177183911204338, 0.040906816720962524, 0.12076317518949509, 0.04317319765686989, -0.07967355847358704, 0.029328804463148117, 0.09720364212989807, -0.019203871488571167, -0.2605144679546356, -0.06618981063365936, -0.014644469134509563, -0.09020574390888214, 0.08537213504314423, 0.05046361684799194, -0.022129623219370842, 0.004671537782996893, -0.005565268453210592, -0.007275679148733616, -0.008651449345052242, 0.04790443181991577, 0.06988821178674698, 0.053793542087078094, 0.09369231015443802, -0.030549511313438416, -0.027574921026825905, 0.06373128294944763, -0.016266506165266037, 0.2204117327928543, -0.06797890365123749, 0.060610268265008926, 0.025512510910630226, 0.08236408233642578, -0.005775300320237875, 0.031072797253727913, 0.021420549601316452, -0.016605859622359276, 0.011050616391003132, -0.04453971982002258, -0.002037788275629282, -0.00864497572183609, -0.016718029975891113, -0.002544519491493702, -0.07331492751836777, 0.03385642170906067, 0.011138711124658585, 0.2765631079673767, 0.035487137734889984, -0.284611314535141, -0.04605555161833763, -0.026503773406147957, -0.028651360422372818, -0.04591040313243866, -0.005629050079733133, 0.08545055985450745, -0.10743239521980286, 0.061792999505996704, -0.03300953656435013, 0.08968554437160492, -0.07400911301374435, 0.0037164974492043257, 0.06542158126831055, 0.142923042178154, -0.018471291288733482, 0.05301152542233467, -0.21800552308559418, 0.22585459053516388, 0.0066343690268695354, 0.12276840209960938, -0.06318739801645279, 0.013110254891216755, 0.016679812222719193, 0.04092568904161453, 0.05855895206332207, -0.0038515652995556593, -0.014002656564116478, -0.16581858694553375, -0.11693862825632095, 0.027765357866883278, 0.11967969685792923, 0.0032486077398061752, 0.07868402451276779, -0.02144247107207775, -0.014906403608620167, 0.0396747961640358, -0.1595315933227539, -0.16270165145397186, -0.12379193305969238, 0.02823607251048088, 0.03448820114135742, -0.044057685881853104, -0.054077114909887314, -0.12470444291830063, -0.039232298731803894, 0.16453835368156433, 0.035076431930065155, -0.05449248477816582, -0.14389315247535706, 0.066850945353508, 0.14511072635650635, -0.041888125240802765, 0.00691293366253376, 0.0384480282664299, 0.11561352759599686, 0.04653283581137657, -0.08428172022104263, 0.034990083426237106, -0.06109831854701042, -0.1877598613500595, -0.056830260902643204, 0.12465117126703262, 0.09073919802904129, 0.053268298506736755, -0.026493046432733536, 0.02750011347234249, 0.003997167572379112, -0.08033448457717896, 0.006247107405215502, 0.08771596848964691, 0.07635447382926941, 0.06108074635267258, -0.06819121539592743, -0.006967482157051563, -0.04995374754071236, -0.02249976061284542, 0.10767213255167007, 0.2037782222032547, -0.07858818024396896, 0.10466939210891724, 0.07101269066333771, -0.06660730391740799, -0.1813579797744751, 0.06008187681436539, 0.12963388860225677, 0.009694674983620644, 0.025962339714169502, -0.1945790946483612, 0.12299437075853348, 0.11637449264526367, -0.00906231440603733, 0.0450182743370533, -0.34279218316078186, -0.12844416499137878, 0.056274645030498505, 0.09172459691762924, 0.03407452628016472, -0.09901013225317001, -0.012945003807544708, -0.0190846286714077, -0.14174602925777435, 0.1401473879814148, -0.0544801726937294, 0.10898800939321518, -0.004719545599073172, 0.09566200524568558, 0.021255727857351303, -0.04220098257064819, 0.10278262197971344, 0.06058437377214432, 0.07064787298440933, -0.04381586238741875, 0.011309950612485409, 0.026816600933670998, -0.06363191455602646, 0.02290087565779686, -0.04024220630526543, 0.07234114408493042, -0.11423646658658981, -0.014378964900970459, -0.08314557373523712, 0.03498023748397827, -0.0470258891582489, -0.043878767639398575, -0.03533090278506279, 0.0474909171462059, 0.0793350413441658, -0.04989689216017723, 0.05730270594358444, 0.0011481177061796188, 0.08706723898649216, 0.05109071731567383, 0.08949291706085205, -0.07241134345531464, -0.08393627405166626, -0.004320865496993065, -0.012367164716124535, 0.03419953212141991, -0.11218557506799698, 0.022140178829431534, 0.13627003133296967, 0.049972500652074814, 0.11657801270484924, 0.039748772978782654, -0.03171136975288391, -0.01968604326248169, 0.032995350658893585, -0.1224730908870697, -0.07667890936136246, 0.0316556915640831, -0.05160103365778923, -0.09848528355360031, 0.0109057342633605, 0.10078628361225128, -0.03410901874303818, -0.022340580821037292, -0.00013555075565818697, 0.023973848670721054, 0.01590842939913273, 0.21333540976047516, 0.026456130668520927, 0.06422258913516998, -0.12091384083032608, 0.12338802218437195, 0.06457487493753433, -0.09908896684646606, 0.04415782168507576, 0.1298801153898239, -0.10387956351041794, -0.010230641812086105, 0.11094801127910614, 0.1498626470565796, -0.0278326328843832, -0.02524591237306595, -0.09082815051078796, -0.10064702481031418, 0.08419262617826462, 0.14142927527427673, 0.03171561285853386, -0.0027477352414280176, -0.054759398102760315, 0.013664370402693748, -0.15279506146907806, 0.060638461261987686, 0.08633929491043091, 0.05642608553171158, -0.07743722200393677, 0.13506051898002625, 0.032538462430238724, 0.035453930497169495, -0.019221382215619087, 0.01559519488364458, -0.07559387385845184, -0.0075123002752661705, -0.08169741928577423, -0.011679033748805523, -0.002456060843542218, -0.0014550605555996299, -0.025930924341082573, -0.04928484559059143, -0.042350318282842636, 0.04086346551775932, -0.076283298432827, -0.05749664828181267, -0.0007092516170814633, 0.03705805912613869, -0.13335131108760834, 0.00587010383605957, 0.001655634376220405, -0.09354188293218613, 0.06905544549226761, 0.041542261838912964, -0.00916486606001854, 0.033602163195610046, -0.16585806012153625, -0.051639724522829056, 0.02889380045235157, 0.017810776829719543, 0.08571425080299377, -0.07641815394163132, 0.00017423488316126168, -0.0197061188519001, 0.044166769832372665, 0.013070347718894482, 0.05295044183731079, -0.11156700551509857, 0.011290603317320347, -0.0625595971941948, -0.03368179500102997, -0.07228302210569382, 0.036864250898361206, 0.1102171540260315, 0.043749306350946426, 0.17331403493881226, -0.07995204627513885, 0.040421366691589355, -0.17826756834983826, -0.03356058895587921, -0.00492100091651082, -0.043163564056158066, -0.07393937557935715, -0.01675092615187168, 0.11734699457883835, -0.05266362056136131, 0.12800289690494537, 0.015007860027253628, 0.0682416558265686, 0.04161416366696358, -0.033677373081445694, -0.06848932057619095, 0.013396055437624454, 0.1441085934638977, 0.06273206323385239, -0.019797369837760925, 0.11145376414060593, -0.00936912465840578, 0.05286112800240517, 0.05334047973155975, 0.23543928563594818, 0.13604037463665009, 0.0411580465734005, 0.06565017998218536, 0.04304759204387665, -0.14207477867603302, -0.10743946582078934, 0.1056281253695488, -0.1007331907749176, 0.12476290762424469, -0.06824374943971634, 0.20919205248355865, 0.046718526631593704, -0.14018838107585907, 0.054197199642658234, -0.03279511258006096, -0.09718036651611328, -0.10699296742677689, 0.00767157506197691, -0.07764140516519547, -0.10908593982458115, 0.03864560276269913, -0.11089880019426346, 0.06591567397117615, 0.09969385713338852, 0.030778752639889717, 0.03155156597495079, 0.10201426595449448, -0.009593711234629154, 0.004230008460581303, 0.062443457543849945, 0.023506782948970795, -0.0038511089514940977, -0.07108660042285919, -0.06649939715862274, 0.028939327225089073, 0.024399902671575546, 0.0836828425526619, -0.0353679321706295, -0.014991797506809235, 0.04896436631679535, -0.01624513790011406, -0.07598420977592468, 0.04334765300154686, -0.003756033256649971, 0.07424551248550415, 0.06145676597952843, 0.0511111356317997, 0.013527355156838894, -0.04783052206039429, 0.2879202961921692, -0.06018925830721855, -0.11890724301338196, -0.11926805973052979, 0.2433442771434784, 0.05929454788565636, -0.024541297927498817, 0.0632113590836525, -0.09979745745658875, -0.054111383855342865, 0.16810916364192963, 0.17400220036506653, -0.06803993880748749, -0.026851220056414604, -0.031375348567962646, -0.010948528535664082, -0.032338667660951614, 0.13908891379833221, 0.11998453736305237, 0.05274169519543648, -0.052968334406614304, -0.027968429028987885, -0.024801574647426605, -0.03739011287689209, -0.06607945263385773, 0.0725693330168724, 0.02668006345629692, -0.006521329283714294, -0.0397363156080246, 0.0672888532280922, -0.01155652292072773, -0.20641663670539856, 0.031626954674720764, -0.17126721143722534, -0.18399356305599213, -0.005311386659741402, 0.10410962998867035, -0.025882145389914513, 0.048959001898765564, 0.0197612252086401, -0.01147228479385376, 0.09813951700925827, -0.03537244349718094, -0.021503303200006485, -0.09887870401144028, 0.05952449142932892, -0.13420617580413818, 0.22563888132572174, -0.005819066893309355, 0.07413896918296814, 0.09286371618509293, 0.032142333686351776, -0.12449746578931808, 0.071816585958004, 0.04134250059723854, -0.1017509177327156, 0.03128782659769058, 0.15915420651435852, -0.043692197650671005, 0.03498134762048721, 0.03821790963411331, -0.12448330968618393, 0.004112954251468182, -0.10055524855852127, -0.03251068294048309, -0.05868850275874138, -0.019375218078494072, -0.028451938182115555, 0.1463722288608551, 0.21213939785957336, -0.006611035671085119, 0.0303301140666008, -0.0791238471865654, -0.010955705307424068, 0.019894765689969063, 0.0824689194560051, -0.029309185221791267, -0.2275806963443756, 0.03629998490214348, 0.013756228610873222, 0.035612836480140686, -0.18619270622730255, -0.06429605931043625, 0.002566365757957101, -0.07444923371076584, -0.0534016378223896, 0.1054806113243103, 0.04190210998058319, 0.04183850437402725, -0.025109494104981422, -0.0928833931684494, -0.008084068074822426, 0.15109869837760925, -0.1884128451347351, -0.03943275287747383 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # albert_distilgpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "albert_distilgpt2_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/albert_distilgpt2_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# albert_distilgpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
[ "# albert_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# albert_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ 62, 43, 6, 12, 8, 3, 118, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# albert_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ -0.08602342009544373, 0.14191675186157227, -0.0035818428732454777, 0.06898482143878937, 0.13063564896583557, 0.046372827142477036, 0.11269187927246094, 0.12615413963794708, -0.09957826137542725, 0.054724860936403275, 0.08051259070634842, 0.04479333385825157, 0.05322207510471344, 0.15099529922008514, -0.03222791478037834, -0.22660817205905914, 0.0025482280179858208, -0.0026386475656181574, -0.07075897604227066, 0.1230790838599205, 0.08736219257116318, -0.09372019022703171, 0.057942379266023636, -0.0106577780097723, -0.13149915635585785, 0.02408081851899624, -0.0417238287627697, -0.05865975469350815, 0.09277012944221497, -0.007129318080842495, 0.09407193958759308, 0.03986151143908501, 0.14895901083946228, -0.2086232453584671, 0.0013273427030071616, 0.08967214077711105, 0.03901257365942001, 0.08961277455091476, 0.06354740262031555, -0.01884073205292225, 0.12018875032663345, -0.17185957729816437, 0.10009165108203888, 0.016011511906981468, -0.07784539461135864, -0.1154312714934349, -0.09865471720695496, 0.07896804809570312, 0.09488741308450699, 0.11655616760253906, 0.0028152598533779383, 0.12331805378198624, -0.11180675029754639, 0.0803738683462143, 0.18236853182315826, -0.20817053318023682, -0.050580885261297226, 0.0598660409450531, 0.018075240775942802, 0.06029659882187843, -0.08711513131856918, -0.014389428310096264, 0.030016988515853882, 0.020530780777335167, 0.10131454467773438, 0.004266857169568539, -0.09339527785778046, 0.003351312829181552, -0.12498622387647629, -0.03697292134165764, 0.10535688698291779, 0.024217866361141205, -0.011693271808326244, -0.10948099941015244, -0.0473325252532959, -0.1482941061258316, -0.011184975504875183, -0.0088697150349617, 0.03522803261876106, -0.06672031432390213, -0.05618330091238022, -0.05209092050790787, -0.07384199649095535, -0.048977918922901154, -0.012945698574185371, 0.1051839292049408, 0.044186074286699295, 0.012043653056025505, -0.039259325712919235, 0.1261558085680008, 0.07115548104047775, -0.1277201920747757, 0.0033015282824635506, 0.012105491012334824, -0.09493831545114517, -0.047911904752254486, -0.02225789614021778, -0.010544680058956146, -0.01693311519920826, 0.13104262948036194, -0.06689866632223129, 0.07259353250265121, 0.02523667924106121, -0.010346335358917713, -0.022711431607604027, 0.14805100858211517, -0.044326089322566986, -0.042942725121974945, -0.021412663161754608, 0.11825606971979141, 0.0031259641982614994, -0.009149412624537945, -0.06756414473056793, -0.0030815093778073788, 0.07089774310588837, 0.06354262679815292, -0.041172586381435394, 0.04444364458322525, -0.02105218730866909, -0.02434566244482994, 0.034847188740968704, -0.14739327132701874, 0.03846057504415512, 0.01876501739025116, -0.09815448522567749, 0.001687020412646234, 0.022947687655687332, -0.012489276006817818, -0.057366758584976196, 0.12115335464477539, -0.0719931423664093, 0.0097108855843544, -0.07910056412220001, -0.06964614242315292, 0.020610840991139412, -0.11005529016256332, -0.04038142040371895, -0.033003516495227814, -0.189909428358078, -0.06249430403113365, 0.025394156575202942, -0.05595117062330246, -0.04847738891839981, -0.05699232593178749, -0.07820384204387665, 0.01652475632727146, -0.010735991410911083, 0.1711498349905014, -0.06093817204236984, 0.06940779834985733, -0.023702347651124, 0.02640637755393982, 0.07760027050971985, 0.047668516635894775, -0.0633159801363945, 0.0012433530064299703, -0.09249117225408554, 0.09334378689527512, -0.08507247269153595, -0.00456869974732399, -0.1052558645606041, -0.10275473445653915, 0.02618195302784443, -0.01041644811630249, 0.04189445450901985, 0.14625506103038788, -0.18360528349876404, -0.038135405629873276, 0.12831375002861023, -0.05896307900547981, -0.0008200535667128861, 0.0626959353685379, -0.052427902817726135, -0.016572333872318268, 0.05982612818479538, 0.12385503947734833, 0.07679494470357895, -0.12449022382497787, -0.012506643310189247, 0.007978045381605625, 0.04515312984585762, 0.010700905695557594, 0.03394660726189613, 0.011832582764327526, 0.05800563097000122, 0.009378567337989807, -0.0675206407904625, 0.004098249599337578, -0.09611912816762924, -0.06923121213912964, -0.036351654678583145, -0.07481583207845688, 0.042743977159261703, 0.03085262142121792, 0.03779495880007744, -0.055032987147569656, -0.11862332373857498, 0.08430097252130508, 0.12338082492351532, -0.057012636214494705, 0.004503163043409586, -0.0774727463722229, -0.04141901433467865, 0.01248854212462902, -0.021092114970088005, -0.19404177367687225, -0.12030956149101257, 0.019530944526195526, -0.059327319264411926, 0.04153010994195938, -0.0004907613038085401, 0.06567247211933136, 0.03743203729391098, -0.047406069934368134, -0.025829652324318886, -0.07637710869312286, -0.003399699227884412, -0.08211281150579453, -0.19279085099697113, -0.048639725893735886, -0.018686702474951744, 0.1799231916666031, -0.2271769940853119, 0.025126280263066292, -0.014681425876915455, 0.1452847570180893, 0.02434815652668476, -0.059553634375333786, 0.02118639647960663, 0.05324086919426918, 0.009582431986927986, -0.09200713783502579, 0.028125988319516182, -0.01216031238436699, -0.07661491632461548, -0.012756510637700558, -0.13354524970054626, -0.02091013826429844, 0.07116778194904327, 0.10270988941192627, -0.11107102036476135, 0.018313664942979813, -0.07695835828781128, -0.04652416333556175, -0.09944114834070206, 0.02221757546067238, 0.20709416270256042, 0.038395754992961884, 0.13163618743419647, -0.04519684612751007, -0.08875206857919693, -0.006275056395679712, 0.015449645929038525, 0.002594328485429287, 0.10935739427804947, 0.06760965287685394, -0.04434590041637421, 0.08091207593679428, 0.01835550367832184, -0.05026978626847267, 0.14063668251037598, -0.052022647112607956, -0.09363527595996857, -0.0009446284966543317, -0.018548792228102684, -0.025516945868730545, 0.10359110683202744, -0.042508602142333984, -0.007225050125271082, 0.03586608171463013, 0.01835394836962223, 0.03436639532446861, -0.1822323501110077, -0.0061624362133443356, 0.02019658125936985, -0.061690136790275574, -0.03615116700530052, -0.02727707102894783, 0.059833403676748276, 0.10390985012054443, -0.0015912216622382402, -0.03163260221481323, 0.01763211376965046, -0.00984532106667757, -0.07881128042936325, 0.17533475160598755, -0.10507248342037201, -0.17883063852787018, -0.10215852409601212, 0.04460665211081505, -0.035183824598789215, -0.05401397868990898, 0.0002305637754034251, -0.10962813347578049, -0.06253387778997421, -0.11303745955228806, -0.016384435817599297, 0.001568026258610189, -0.011550038121640682, 0.05262640863656998, 0.01664375700056553, 0.049889713525772095, -0.12726515531539917, 0.013700527139008045, -0.03808445855975151, -0.08196805417537689, 0.027888108044862747, 0.062477413564920425, 0.05817744508385658, 0.14949364960193634, -0.01897505298256874, 0.03230448067188263, -0.02026614546775818, 0.17760691046714783, -0.09722601622343063, 0.00996890477836132, 0.10950346291065216, 0.012249608524143696, 0.04081963002681732, 0.11100152879953384, 0.034339677542448044, -0.06378205120563507, 0.0175296813249588, 0.09132722020149231, -0.026017339900135994, -0.2693089544773102, -0.06538517028093338, -0.014754445292055607, -0.06543736904859543, 0.09231306612491608, 0.05653480067849159, -0.005876727402210236, 0.009807699359953403, -0.01640980876982212, -0.014401474967598915, 0.0006456906558014452, 0.052791886031627655, 0.0825318917632103, 0.042502228170633316, 0.0878445953130722, -0.03402205556631088, -0.030407344922423363, 0.06197088584303856, 0.0005546418251469731, 0.2320972979068756, -0.06063206121325493, 0.07820235937833786, 0.012297112494707108, 0.0921446904540062, -0.010220721364021301, 0.03683886677026749, 0.033069223165512085, -0.008244131691753864, 0.01871459372341633, -0.04485617205500603, -0.005472524557262659, 0.006686551962047815, -0.006793221924453974, -0.0029520399402827024, -0.06921929866075516, 0.033609334379434586, 0.006524224765598774, 0.29899853467941284, 0.04493061453104019, -0.2899249196052551, -0.05394546315073967, -0.018787004053592682, -0.03864408656954765, -0.04979882389307022, -0.009384880773723125, 0.10562705993652344, -0.12766698002815247, 0.0718129426240921, -0.04417029023170471, 0.08326303958892822, -0.07483679801225662, -0.0028983955271542072, 0.05663932487368584, 0.11193662136793137, -0.010572534054517746, 0.07019773125648499, -0.21007804572582245, 0.2334185093641281, 0.005739678628742695, 0.1116369366645813, -0.07229077816009521, 0.02242223545908928, 0.015679936856031418, 0.04696144908666611, 0.08799748122692108, -0.005523769184947014, -0.0172407403588295, -0.1577400118112564, -0.13095097243785858, 0.02155829221010208, 0.11382140219211578, -0.020401649177074432, 0.08379167318344116, -0.018368884921073914, -0.01838173158466816, 0.028998857364058495, -0.15061277151107788, -0.14424648880958557, -0.12815557420253754, 0.03827294334769249, 0.028003131970763206, -0.034022267907857895, -0.059565890580415726, -0.12302977591753006, -0.023299697786569595, 0.18460670113563538, 0.002808196237310767, -0.06728005409240723, -0.14746016263961792, 0.0665888711810112, 0.14694632589817047, -0.041852183640003204, -0.00010634717182256281, 0.04484960809350014, 0.12477832287549973, 0.03931562229990959, -0.08620048314332962, 0.020428413525223732, -0.06085110828280449, -0.19255748391151428, -0.04499762877821922, 0.14194028079509735, 0.08501319587230682, 0.05709978565573692, -0.01785549521446228, 0.0242687426507473, 0.011649015359580517, -0.07604410499334335, 0.01015156414359808, 0.09106236696243286, 0.08010994642972946, 0.07197076082229614, -0.07166063785552979, -0.008458797819912434, -0.04846145585179329, -0.021181685850024223, 0.10852882266044617, 0.1752283126115799, -0.08406136184930801, 0.10580576956272125, 0.056959521025419235, -0.07088422775268555, -0.1776546984910965, 0.04553194344043732, 0.1369810551404953, 0.021642379462718964, 0.03510626032948494, -0.20561428368091583, 0.11307179927825928, 0.12700320780277252, -0.012021155096590519, 0.03700558468699455, -0.34696975350379944, -0.11898176372051239, 0.06148383393883705, 0.07119826972484589, 0.01529400423169136, -0.10996680706739426, -0.027266476303339005, 0.0005160790751688182, -0.1342175006866455, 0.13683976233005524, -0.052549924701452255, 0.10576868057250977, -0.011222539469599724, 0.10903123766183853, 0.02917686104774475, -0.04740364849567413, 0.12045592069625854, 0.0758805051445961, 0.06110333278775215, -0.044021476060152054, 0.010671661235392094, 0.051803428679704666, -0.07507678121328354, 0.06053173169493675, -0.045935146510601044, 0.06476239860057831, -0.1510060727596283, -0.019928274676203728, -0.08897648751735687, 0.04498129338026047, -0.04556999355554581, -0.03683977201581001, -0.028031932190060616, 0.05316498875617981, 0.06840281933546066, -0.04553694650530815, 0.06583163142204285, 0.0013111460721120238, 0.09071604907512665, 0.08385471254587173, 0.09915192425251007, -0.03192511945962906, -0.11002831906080246, -0.021279877051711082, -0.011084064841270447, 0.0358753576874733, -0.10984145104885101, 0.01776890642940998, 0.1399899125099182, 0.057989366352558136, 0.1338064819574356, 0.028469545766711235, -0.03889116644859314, -0.02017781138420105, 0.03526097163558006, -0.12443256378173828, -0.10403282195329666, 0.009477558545768261, -0.05974964052438736, -0.11682657152414322, 0.007570063695311546, 0.10222692787647247, -0.038207780569791794, -0.023476479575037956, -0.0010193557245656848, 0.03255366533994675, 0.011942675337195396, 0.2133103460073471, 0.02370341680943966, 0.07714705914258957, -0.11269374936819077, 0.11593838036060333, 0.06738795340061188, -0.10751534253358841, 0.045831333845853806, 0.13088521361351013, -0.09541172534227371, -0.01760338805615902, 0.09348396211862564, 0.13191446661949158, -0.034165482968091965, -0.03290943056344986, -0.09354736655950546, -0.10558276623487473, 0.07829851657152176, 0.12160474061965942, 0.037032973021268845, 0.0113541129976511, -0.046128906309604645, 0.00430450402200222, -0.14454036951065063, 0.06925973296165466, 0.09363563358783722, 0.05371948331594467, -0.09090730547904968, 0.14338506758213043, 0.024488728493452072, 0.04882987216114998, -0.016982372850179672, 0.0132227074354887, -0.0646781399846077, -0.006091552320867777, -0.08131802827119827, -0.00558877969160676, -0.007194269448518753, -0.0031622136011719704, -0.01764088124036789, -0.04376790300011635, -0.03334246948361397, 0.04205876588821411, -0.07950713485479355, -0.057224053889513016, -0.0071401530876755714, 0.046770915389060974, -0.13183346390724182, -0.00023384099768009037, 0.0044448100961744785, -0.10195986926555634, 0.07326105237007141, 0.05100817233324051, -0.006287507712841034, 0.029827091842889786, -0.1226339116692543, -0.03783482313156128, 0.02946210838854313, 0.022508233785629272, 0.06845386326313019, -0.07392381131649017, 0.005326662678271532, -0.020884912461042404, 0.02707616612315178, 0.024730177596211433, 0.03387944772839546, -0.11315794289112091, 0.007346836384385824, -0.06385137885808945, -0.034863147884607315, -0.07751630246639252, 0.047813452780246735, 0.10887829214334488, 0.04221627488732338, 0.16576474905014038, -0.08201843500137329, 0.048055876046419144, -0.18622538447380066, -0.03113015927374363, -0.0019207862205803394, -0.032840702682733536, -0.053683992475271225, -0.015130534768104553, 0.11159432679414749, -0.04604149982333183, 0.11802869290113449, 0.0107226287946105, 0.06334564834833145, 0.03995019569993019, -0.036948736757040024, -0.05673903971910477, 0.013319532386958599, 0.13573333621025085, 0.05678989365696907, -0.021798502653837204, 0.11060407012701035, -0.02470318041741848, 0.035705067217350006, 0.045492712408304214, 0.22747519612312317, 0.13245926797389984, 0.010115601122379303, 0.06833074241876602, 0.05840153247117996, -0.14031195640563965, -0.12133345752954483, 0.09687259048223495, -0.08295094221830368, 0.12784312665462494, -0.06397287547588348, 0.20424914360046387, 0.061010900884866714, -0.15309302508831024, 0.05985202640295029, -0.027086298912763596, -0.10466697067022324, -0.10389337688684464, -0.0227165799587965, -0.07755724340677261, -0.10049692541360855, 0.037646107375621796, -0.11596186459064484, 0.060895178467035294, 0.10265535861253738, 0.03320074826478958, 0.02338247187435627, 0.11250333487987518, -0.008621206507086754, -0.00463081942871213, 0.07547519356012344, 0.023354779928922653, 0.002615571254864335, -0.08169993758201599, -0.05616336315870285, 0.038518451154232025, 0.025699613615870476, 0.08461037278175354, -0.03660157322883606, 0.002836512168869376, 0.033821169286966324, -0.017932701855897903, -0.07596004009246826, 0.034619156271219254, -0.0008229986415244639, 0.0728621706366539, 0.06376402080059052, 0.05066022276878357, 0.0036893293727189302, -0.05395394563674927, 0.27465999126434326, -0.0658416673541069, -0.10714966058731079, -0.13683108985424042, 0.23266693949699402, 0.046956196427345276, -0.027290118858218193, 0.06990799307823181, -0.10417432337999344, -0.039041146636009216, 0.1525854766368866, 0.16781292855739594, -0.03459109738469124, -0.022000035271048546, -0.03363603353500366, -0.009020617231726646, -0.018680797889828682, 0.112834133207798, 0.10050094872713089, 0.07204781472682953, -0.0551578551530838, -0.020151685923337936, -0.010065844282507896, -0.03884382173418999, -0.07582004368305206, 0.07920252531766891, 0.022299980744719505, -0.005268396344035864, -0.03796869143843651, 0.07617954164743423, -0.010582711547613144, -0.19634081423282623, 0.04877251386642456, -0.17694666981697083, -0.1841145157814026, -0.010708898305892944, 0.08990706503391266, -0.024578751996159554, 0.04833994433283806, 0.0164874829351902, -0.02027021162211895, 0.11369775980710983, -0.030337611213326454, -0.05100835859775543, -0.09611566364765167, 0.06947658956050873, -0.11607497185468674, 0.2247350513935089, -0.0017022636020556092, 0.0783192366361618, 0.08557648211717606, 0.012509152293205261, -0.1291607916355133, 0.056809715926647186, 0.04200132191181183, -0.08037176728248596, 0.027804631739854813, 0.17988555133342743, -0.044209450483322144, 0.04119184985756874, 0.03016921877861023, -0.1259654015302658, -0.014218038879334927, -0.07282789051532745, -0.026883214712142944, -0.0646975040435791, -0.010894018225371838, -0.023039644584059715, 0.1459933966398239, 0.22342169284820557, -0.011956734582781792, 0.022486615926027298, -0.0767211839556694, -0.004501464311033487, 0.030781131237745285, 0.08189792186021805, -0.014463488943874836, -0.21046721935272217, 0.027073506265878677, -0.021247869357466698, 0.03885148465633392, -0.19443897902965546, -0.07546992599964142, 0.008162458427250385, -0.06647475063800812, -0.049771953374147415, 0.10269183665513992, 0.055037084966897964, 0.04313286393880844, -0.018360072746872902, -0.07535748928785324, -0.008798477239906788, 0.14524266123771667, -0.19201168417930603, -0.03299642726778984 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # albert_gpt2_Full_summarization_cnndm This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "albert_gpt2_Full_summarization_cnndm", "results": []}]}
text2text-generation
Ayham/albert_gpt2_Full_summarization_cnndm
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# albert_gpt2_Full_summarization_cnndm This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
[ "# albert_gpt2_Full_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# albert_gpt2_Full_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ 62, 41, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# albert_gpt2_Full_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ -0.08498145639896393, 0.1349177360534668, -0.003717264160513878, 0.0712609589099884, 0.13565534353256226, 0.04321669787168503, 0.10237523913383484, 0.11523386836051941, -0.12123723328113556, 0.05477811396121979, 0.07747439295053482, 0.0609186589717865, 0.047570426017045975, 0.14058417081832886, -0.03193008899688721, -0.22000360488891602, 0.0005808659479953349, -0.001730420277453959, -0.0811031386256218, 0.11239766329526901, 0.0885457694530487, -0.10454769432544708, 0.0593496672809124, -0.005611451342701912, -0.13896127045154572, 0.018216075375676155, -0.04127374663949013, -0.04833683744072914, 0.09303600341081619, -0.003572849091142416, 0.09843224287033081, 0.03313450515270233, 0.14592468738555908, -0.21011869609355927, 0.0026300179306417704, 0.09441739320755005, 0.0439484603703022, 0.09249351918697357, 0.05962236225605011, -0.005201692692935467, 0.12641334533691406, -0.1646653562784195, 0.10354040563106537, 0.01455026213079691, -0.07691246271133423, -0.11845703423023224, -0.09312848746776581, 0.06293030828237534, 0.0942634791135788, 0.11682865023612976, 0.005077901761978865, 0.11985234171152115, -0.10827433317899704, 0.0808231383562088, 0.17082545161247253, -0.21606449782848358, -0.052812330424785614, 0.06638723611831665, 0.03238792344927788, 0.06479585915803909, -0.08903983980417252, -0.006100947968661785, 0.031199753284454346, 0.021010149270296097, 0.10078078508377075, 0.0034887981601059437, -0.10084942728281021, 0.0031759669072926044, -0.12294932454824448, -0.023309214040637016, 0.08834994584321976, 0.02263367548584938, -0.009924297221004963, -0.10158071666955948, -0.053111497312784195, -0.1559922695159912, -0.021665114909410477, -0.021231045946478844, 0.042600132524967194, -0.05902071297168732, -0.06288968771696091, -0.05125534161925316, -0.06784687936306, -0.056202568113803864, -0.02370261214673519, 0.11980647593736649, 0.04513723775744438, 0.007546791806817055, -0.045153696089982986, 0.1260317713022232, 0.060423169285058975, -0.12357079237699509, 0.009132315404713154, 0.011566630564630032, -0.09773221611976624, -0.05434126779437065, -0.030597852542996407, -0.01927088014781475, -0.010719681158661842, 0.1345691978931427, -0.0763666182756424, 0.07777296006679535, 0.023349951952695847, -0.012453313916921616, -0.016278797760605812, 0.14637912809848785, -0.05561736226081848, -0.037264253944158554, -0.014512340538203716, 0.1053130179643631, 0.0074542006477713585, -0.014700030907988548, -0.07365838438272476, -0.00787077471613884, 0.06721808761358261, 0.06344740837812424, -0.04591935873031616, 0.04226665198802948, -0.027661718428134918, -0.02240666374564171, 0.036095790565013885, -0.1325034648180008, 0.04403674975037575, 0.012828707695007324, -0.09716670215129852, -0.00846492126584053, 0.016381723806262016, 0.002195740584284067, -0.046113818883895874, 0.13320176303386688, -0.074956513941288, 0.007006555330008268, -0.08227396011352539, -0.07232342660427094, 0.012023611925542355, -0.10645734518766403, -0.030372897163033485, -0.037404365837574005, -0.18258357048034668, -0.05264882743358612, 0.033193621784448624, -0.05923888087272644, -0.041693635284900665, -0.05615903437137604, -0.07282573729753494, 0.02049821801483631, -0.008921400643885136, 0.1848035752773285, -0.06776929646730423, 0.06220525503158569, -0.027596017345786095, 0.026172852143645287, 0.057319071143865585, 0.04108811542391777, -0.06191571056842804, -0.001523651764728129, -0.10001157224178314, 0.08933378010988235, -0.07515358179807663, 0.007880177348852158, -0.10538985580205917, -0.0977042019367218, 0.01935327611863613, -0.010852590203285217, 0.057152725756168365, 0.13244742155075073, -0.19895558059215546, -0.0319247841835022, 0.11738675832748413, -0.04394249990582466, -0.014053542166948318, 0.06092268228530884, -0.05663766339421272, -0.010775892995297909, 0.06352102011442184, 0.12475818395614624, 0.06655372679233551, -0.12783943116664886, -0.0035207243636250496, 0.007854437455534935, 0.04058558866381645, 0.016853678971529007, 0.026446592062711716, 0.006222577765583992, 0.05368886515498161, 0.003544426755979657, -0.039465274661779404, 0.004205996170639992, -0.0978010818362236, -0.07085847854614258, -0.03769560530781746, -0.07824026048183441, 0.021876318380236626, 0.025933653116226196, 0.03421376273036003, -0.056909605860710144, -0.11972864717245102, 0.08313805609941483, 0.12595151364803314, -0.05514506995677948, 0.017701048403978348, -0.07275944948196411, -0.03105219267308712, 0.009193756617605686, -0.025071941316127777, -0.19213098287582397, -0.10832520574331284, 0.021916942670941353, -0.06393897533416748, 0.03735871613025665, -0.0056318617425858974, 0.0636986643075943, 0.04517820104956627, -0.04622562229633331, -0.01538860984146595, -0.06966006755828857, 0.00003446360642556101, -0.09277381747961044, -0.21290040016174316, -0.031316570937633514, -0.015111793763935566, 0.17653293907642365, -0.21908269822597504, 0.013758056797087193, -0.02408141829073429, 0.1471843719482422, 0.02167171984910965, -0.058821313083171844, 0.0140567347407341, 0.05046652629971504, 0.004209544509649277, -0.0938667580485344, 0.03036249801516533, -0.005305514670908451, -0.08564554899930954, -0.025335028767585754, -0.13583111763000488, -0.02261217124760151, 0.08103187382221222, 0.08084949105978012, -0.10605420917272568, 0.019018730148673058, -0.06941507756710052, -0.044641975313425064, -0.08328203856945038, 0.03352879732847214, 0.22164656221866608, 0.029650693759322166, 0.13420096039772034, -0.05151268467307091, -0.08566565066576004, -0.006325362715870142, 0.02486642450094223, 0.011293558403849602, 0.10206294804811478, 0.08678306639194489, -0.04222366213798523, 0.07140744477510452, 0.02441740594804287, -0.0489245280623436, 0.13661512732505798, -0.042107462882995605, -0.08692098408937454, -0.003056902904063463, -0.025080019608139992, -0.02439802885055542, 0.10726740211248398, -0.05672215670347214, 0.007444052025675774, 0.03159560635685921, 0.02187463454902172, 0.038719236850738525, -0.18386030197143555, -0.0007127264980226755, 0.022615043446421623, -0.056239139288663864, -0.035841651260852814, -0.02503744326531887, 0.051064204424619675, 0.10263587534427643, 0.010480880737304688, -0.02947305329144001, 0.008585421368479729, -0.005791196133941412, -0.06761790066957474, 0.17809055745601654, -0.11471077799797058, -0.1642884761095047, -0.08263323456048965, 0.036529671400785446, -0.03664837032556534, -0.048641640692949295, -0.0013027797685936093, -0.11419660598039627, -0.06915789842605591, -0.11040148884057999, -0.02005956694483757, -0.005083517171442509, -0.004656567238271236, 0.05498787760734558, 0.010636081919074059, 0.05105586349964142, -0.13665680587291718, 0.015370911918580532, -0.04603941738605499, -0.08158503472805023, 0.014400016516447067, 0.08060672879219055, 0.06266161054372787, 0.15714089572429657, -0.016681108623743057, 0.0267561636865139, -0.01784248650074005, 0.18919768929481506, -0.09723372012376785, 0.007752160541713238, 0.09590686112642288, 0.011010349728167057, 0.034711502492427826, 0.11978142708539963, 0.03931457921862602, -0.08011771738529205, 0.024403443560004234, 0.09035839140415192, -0.019311096519231796, -0.26177242398262024, -0.06538868695497513, -0.019290124997496605, -0.07872650772333145, 0.08765508234500885, 0.05082526430487633, -0.003804053645581007, 0.011699995025992393, -0.009096551686525345, 0.0028013859409838915, -0.005505891516804695, 0.05522030219435692, 0.0742790699005127, 0.05131199583411217, 0.09428351372480392, -0.03061230294406414, -0.029070185497403145, 0.06537306308746338, -0.008826297707855701, 0.226088285446167, -0.06045953929424286, 0.08447061479091644, 0.02060752734541893, 0.09159555286169052, -0.011869899928569794, 0.028157450258731842, 0.025176862254738808, -0.013323089107871056, 0.014836715534329414, -0.04931158572435379, -0.002012180397287011, 0.0012110735988244414, -0.010499708354473114, 0.0010094577446579933, -0.0670752003788948, 0.03473273292183876, 0.016379952430725098, 0.2707110643386841, 0.04283561557531357, -0.2897208034992218, -0.05228624865412712, -0.015416759997606277, -0.029882237315177917, -0.058579958975315094, -0.004493578802794218, 0.09904418885707855, -0.11199761182069778, 0.07164713740348816, -0.04236233979463577, 0.08595908433198929, -0.07002726942300797, -0.0005670935497619212, 0.06691848486661911, 0.13682760298252106, -0.015652572736144066, 0.06267468631267548, -0.21976804733276367, 0.21650660037994385, 0.011661696247756481, 0.12154124677181244, -0.06706074625253677, 0.024260856211185455, 0.012848264537751675, 0.055306632071733475, 0.07549741119146347, -0.0033856546506285667, -0.01419336162507534, -0.15372887253761292, -0.12282358109951019, 0.02982538752257824, 0.11780589818954468, -0.009630350396037102, 0.07780351489782333, -0.02016421966254711, -0.014557840302586555, 0.04103175923228264, -0.14867864549160004, -0.16393828392028809, -0.12650781869888306, 0.03306609392166138, 0.02644571289420128, -0.039627715945243835, -0.05454490706324577, -0.11622954905033112, -0.03870856389403343, 0.1818835735321045, 0.014419738203287125, -0.06039119139313698, -0.13918016850948334, 0.06458040326833725, 0.14983074367046356, -0.03738851100206375, 0.01565205492079258, 0.03529787063598633, 0.11744189262390137, 0.045814454555511475, -0.08044660091400146, 0.03843133896589279, -0.06122279912233353, -0.1861671507358551, -0.05294022709131241, 0.13158760964870453, 0.08354420214891434, 0.05259260907769203, -0.019900605082511902, 0.028944287449121475, 0.0022640733513981104, -0.08226148039102554, 0.0036548201460391283, 0.08839813619852066, 0.08770227432250977, 0.06713081151247025, -0.06695915758609772, 0.006662582978606224, -0.04049431160092354, -0.026473980396986008, 0.11003279685974121, 0.19183799624443054, -0.08256280422210693, 0.09910820424556732, 0.05987858027219772, -0.07522286474704742, -0.17569155991077423, 0.06243237480521202, 0.12622766196727753, 0.01704786717891693, 0.028074542060494423, -0.20538821816444397, 0.11913763731718063, 0.12418811768293381, -0.010262257419526577, 0.05464782193303108, -0.34790968894958496, -0.13423924148082733, 0.05710756406188011, 0.0848730280995369, 0.014704267494380474, -0.10555471479892731, -0.01980321854352951, -0.020999815315008163, -0.14170069992542267, 0.14134939014911652, -0.07058817893266678, 0.10906659066677094, -0.007048857398331165, 0.09172296524047852, 0.027854938060045242, -0.047010261565446854, 0.12511935830116272, 0.060204386711120605, 0.0719742625951767, -0.04683603718876839, 0.01924857683479786, 0.044522788375616074, -0.06839442998170853, 0.03847511485219002, -0.04516545310616493, 0.06858626753091812, -0.125113382935524, -0.019061841070652008, -0.08091600984334946, 0.042210303246974945, -0.0504230260848999, -0.047497719526290894, -0.04332834854722023, 0.04994375631213188, 0.07398857176303864, -0.045791395008563995, 0.06989391148090363, 0.004659694153815508, 0.08417779952287674, 0.05911218002438545, 0.10143033415079117, -0.05038866773247719, -0.0834478884935379, -0.001784399850293994, -0.012116249650716782, 0.039866797626018524, -0.12056711316108704, 0.027040647342801094, 0.13997694849967957, 0.04821163788437843, 0.122928686439991, 0.0345468744635582, -0.033184733241796494, -0.022437069565057755, 0.0370696522295475, -0.12327062338590622, -0.09906913340091705, 0.020642738789319992, -0.0726979449391365, -0.10023296624422073, 0.013041692785918713, 0.10122516751289368, -0.03902054578065872, -0.015236173756420612, 0.0024538564030081034, 0.02789398841559887, 0.004503571428358555, 0.2084626853466034, 0.02120385505259037, 0.06448820233345032, -0.1139320656657219, 0.12454438209533691, 0.06804299354553223, -0.09236282110214233, 0.04917030408978462, 0.11346396803855896, -0.0967574417591095, -0.011640493758022785, 0.09656355530023575, 0.15219727158546448, -0.026997311040759087, -0.03341434523463249, -0.09260542690753937, -0.10154163837432861, 0.07611764967441559, 0.11862293630838394, 0.02928532101213932, 0.004268433898687363, -0.05121185630559921, 0.017761286348104477, -0.15585999190807343, 0.06858888268470764, 0.08179653435945511, 0.05946246162056923, -0.09523344784975052, 0.14804722368717194, 0.02613401599228382, 0.033468518406152725, -0.019075876101851463, 0.0175931379199028, -0.06681478768587112, -0.006745233200490475, -0.10166990756988525, -0.012851927429437637, -0.007774988189339638, -0.001969141187146306, -0.02067483775317669, -0.04037994146347046, -0.044342704117298126, 0.03910813480615616, -0.07437039911746979, -0.05588053539395332, -0.002501236740499735, 0.03565627709031105, -0.13081111013889313, 0.0003241464146412909, -0.0005803221138194203, -0.09420348703861237, 0.07564027607440948, 0.043952684849500656, 0.0022168264258652925, 0.03534359857439995, -0.14795725047588348, -0.034677859395742416, 0.030808638781309128, 0.023184500634670258, 0.07173401117324829, -0.07511142641305923, -0.0023375635500997305, -0.01949845626950264, 0.03906603530049324, 0.0164968129247427, 0.0399097241461277, -0.109901063144207, 0.0020723924972116947, -0.05996356159448624, -0.031568631529808044, -0.0692879855632782, 0.04555080085992813, 0.10255058109760284, 0.041727419942617416, 0.16931995749473572, -0.07695213705301285, 0.03653864562511444, -0.18076039850711823, -0.03255648910999298, -0.003212286625057459, -0.03606947511434555, -0.06437312066555023, -0.01842588186264038, 0.11247279495000839, -0.05723843351006508, 0.13670088350772858, 0.006877753883600235, 0.06238605082035065, 0.03918013349175453, -0.0260965283960104, -0.05625971034169197, 0.01724466308951378, 0.15323568880558014, 0.06488826870918274, -0.020017798990011215, 0.10805269330739975, -0.007641464937478304, 0.057540733367204666, 0.05537426099181175, 0.23145273327827454, 0.1296510547399521, 0.020825538784265518, 0.06597213447093964, 0.048867642879486084, -0.14378149807453156, -0.11385241895914078, 0.11786279082298279, -0.09612412750720978, 0.128961980342865, -0.06279939413070679, 0.1899406611919403, 0.05394371598958969, -0.15407349169254303, 0.05355415493249893, -0.03264398127794266, -0.10883230715990067, -0.10913512855768204, -0.013301381841301918, -0.07842189073562622, -0.10663323849439621, 0.03255603089928627, -0.11569878458976746, 0.058138713240623474, 0.09670528024435043, 0.03559959679841995, 0.02111024037003517, 0.11812280863523483, -0.004804154392331839, 0.00000670340114083956, 0.06677234172821045, 0.02720126509666443, -0.0007599075906910002, -0.0722738727927208, -0.06865503638982773, 0.03298304229974747, 0.025254951789975166, 0.08323502540588379, -0.041600801050662994, -0.0084264250472188, 0.04444020986557007, -0.008792777545750141, -0.07057558745145798, 0.03738945722579956, 0.002548223827034235, 0.06528393924236298, 0.0537109412252903, 0.05016608163714409, 0.010579785332083702, -0.048464857041835785, 0.2795812487602234, -0.06316689401865005, -0.11121691763401031, -0.12516646087169647, 0.22151824831962585, 0.04826650023460388, -0.023001471534371376, 0.07214998453855515, -0.10237366706132889, -0.0464291051030159, 0.15649716556072235, 0.1739111691713333, -0.06560923159122467, -0.02566494233906269, -0.0251026451587677, -0.010666165500879288, -0.03037971258163452, 0.12587063014507294, 0.10866087675094604, 0.05101729556918144, -0.05021217465400696, -0.025777319446206093, -0.011361273936927319, -0.03641127422451973, -0.06989752501249313, 0.06959907710552216, 0.022755032405257225, -0.006165598519146442, -0.032800476998090744, 0.06807707995176315, -0.01712876372039318, -0.19577021896839142, 0.04022249951958656, -0.1727861911058426, -0.18816038966178894, -0.01266532950103283, 0.09882402420043945, -0.027630334720015526, 0.042759839445352554, 0.015653660520911217, -0.01561581902205944, 0.1079169362783432, -0.029636066406965256, -0.03625980019569397, -0.10296403616666794, 0.06696385890245438, -0.11285100132226944, 0.23073191940784454, -0.002279717708006501, 0.07211969047784805, 0.09100893139839172, 0.021912207826972008, -0.12317317724227905, 0.058293189853429794, 0.048257820308208466, -0.08911633491516113, 0.03360430896282196, 0.16214942932128906, -0.04676432907581329, 0.03747265040874481, 0.035614434629678726, -0.1140487790107727, -0.003854954382404685, -0.08165664225816727, -0.033376146107912064, -0.0669620931148529, -0.017887461930513382, -0.0357092022895813, 0.149219810962677, 0.21696244180202484, -0.01006028801202774, 0.027613529935479164, -0.07786563038825989, 0.00020698450680356473, 0.025595085695385933, 0.0937829241156578, -0.0249655582010746, -0.2233273983001709, 0.03404936566948891, 0.008454089984297752, 0.04052828997373581, -0.18646366894245148, -0.06782575696706772, 0.012374844402074814, -0.06953375786542892, -0.05434301495552063, 0.10713670402765274, 0.05105758458375931, 0.04197236895561218, -0.024396143853664398, -0.08596577495336533, -0.012859867885708809, 0.14590297639369965, -0.18295416235923767, -0.03830970823764801 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # albert_large_gpt2_summarization_cnndm This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "albert_large_gpt2_summarization_cnndm", "results": []}]}
text2text-generation
Ayham/albert_gpt2_summarization_cnndm
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# albert_large_gpt2_summarization_cnndm This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
[ "# albert_large_gpt2_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# albert_large_gpt2_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ 62, 42, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# albert_large_gpt2_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ -0.08849732577800751, 0.14482295513153076, -0.0037294630892574787, 0.07549097388982773, 0.13146986067295074, 0.04344387352466583, 0.09911063313484192, 0.12568122148513794, -0.10343892872333527, 0.06393788754940033, 0.08082947134971619, 0.04245892912149429, 0.055855151265859604, 0.14006571471691132, -0.03170057386159897, -0.22508373856544495, 0.0013339613797143102, -0.0009233503369614482, -0.06962160766124725, 0.11055347323417664, 0.0908469408750534, -0.10230278968811035, 0.06395772844552994, -0.004702265374362469, -0.13616982102394104, 0.017568716779351234, -0.04095802456140518, -0.049302320927381516, 0.08683671057224274, -0.003142488654702902, 0.1013842299580574, 0.028878768905997276, 0.14489437639713287, -0.21118174493312836, 0.0007116267224773765, 0.08952180296182632, 0.04289738088846207, 0.09035487473011017, 0.06381840258836746, -0.00488419272005558, 0.12151481211185455, -0.16431112587451935, 0.09817807376384735, 0.017822396010160446, -0.07601842284202576, -0.10967342555522919, -0.09263362735509872, 0.06263431161642075, 0.09783587604761124, 0.11492502689361572, 0.007278557401150465, 0.10767435282468796, -0.10227055102586746, 0.08048693835735321, 0.1743512600660324, -0.2150374799966812, -0.055166684091091156, 0.07189301401376724, 0.03697981685400009, 0.06500966846942902, -0.09427989274263382, -0.011314302682876587, 0.030093131586909294, 0.025549231097102165, 0.10815680027008057, -0.0003700426605064422, -0.09752976149320602, 0.002085841493681073, -0.12138745933771133, -0.0310873594135046, 0.09642896056175232, 0.023386942222714424, -0.0097928736358881, -0.10873545706272125, -0.05540342628955841, -0.15866494178771973, -0.017818208783864975, -0.01597696729004383, 0.040087878704071045, -0.054971564561128616, -0.047299426048994064, -0.05875977873802185, -0.06994537264108658, -0.06136679649353027, -0.019279025495052338, 0.10407751798629761, 0.03990402817726135, 0.007809240370988846, -0.04343387857079506, 0.12760847806930542, 0.052691005170345306, -0.12702970206737518, -0.00013154945918358862, 0.012939502485096455, -0.09578733891248703, -0.05179153010249138, -0.026965321972966194, -0.02170448563992977, -0.008178910240530968, 0.1294536143541336, -0.07483960688114166, 0.08089208602905273, 0.019249549135565758, -0.008993280120193958, -0.02476317249238491, 0.15608981251716614, -0.047868818044662476, -0.045066364109516144, -0.013783561065793037, 0.11300003528594971, 0.005901286844164133, -0.011396723799407482, -0.07313868403434753, -0.018672285601496696, 0.06738817691802979, 0.06189602613449097, -0.049896083772182465, 0.036206163465976715, -0.030168648809194565, -0.021783825010061264, 0.035836800932884216, -0.13462722301483154, 0.04485313594341278, 0.01262305211275816, -0.09509730339050293, -0.004381366539746523, 0.01815168187022209, -0.0048585073091089725, -0.04688410460948944, 0.1229177787899971, -0.07375910133123398, 0.004817192442715168, -0.08613631874322891, -0.06968625634908676, 0.016390983015298843, -0.10923641920089722, -0.036158621311187744, -0.03423271328210831, -0.1917657107114792, -0.05589566379785538, 0.029754038900136948, -0.05752480402588844, -0.03849756345152855, -0.058521758764982224, -0.07289120554924011, 0.019775673747062683, -0.007693828083574772, 0.17121949791908264, -0.06290559470653534, 0.06485368311405182, -0.025059286504983902, 0.03118622861802578, 0.053613096475601196, 0.04155539348721504, -0.0603848472237587, 0.006074034608900547, -0.1058429479598999, 0.08902673423290253, -0.07726490497589111, 0.0033045271411538124, -0.10551325976848602, -0.09778927266597748, 0.01211673952639103, -0.011239934712648392, 0.0537494458258152, 0.1367509514093399, -0.19522641599178314, -0.02803289331495762, 0.13067103922367096, -0.047852326184511185, -0.012433838099241257, 0.06379254162311554, -0.052267879247665405, -0.012051818892359734, 0.06017407029867172, 0.12868061661720276, 0.08449000120162964, -0.12731312215328217, -0.00910853035748005, 0.011448429897427559, 0.03996136412024498, 0.012106579728424549, 0.02978622354567051, 0.0029484739061444998, 0.05568404495716095, 0.00784967839717865, -0.04906322434544563, 0.005752757657319307, -0.08948305249214172, -0.07006524503231049, -0.040242359042167664, -0.0760648250579834, 0.021565837785601616, 0.023330802097916603, 0.03161619231104851, -0.055006805807352066, -0.11375780403614044, 0.07971557229757309, 0.1274293214082718, -0.05849209427833557, 0.011316688731312752, -0.07277707755565643, -0.029853053390979767, 0.007656514644622803, -0.024441109970211983, -0.19273483753204346, -0.11078891158103943, 0.02652755379676819, -0.06945392489433289, 0.03496440872550011, -0.0068137855269014835, 0.06616531312465668, 0.04563644155859947, -0.0455692820250988, -0.021150363609194756, -0.07085368782281876, -0.006318950559943914, -0.09254223853349686, -0.19610093533992767, -0.03358669579029083, -0.012914593331515789, 0.17596974968910217, -0.22701524198055267, 0.0188266783952713, -0.009488006122410297, 0.14878858625888824, 0.025121647864580154, -0.05838676914572716, 0.018735811114311218, 0.050291117280721664, 0.002777875866740942, -0.09664090722799301, 0.0274321548640728, -0.007917444221675396, -0.08618512004613876, -0.027037518098950386, -0.1336449682712555, -0.016668733209371567, 0.07574764639139175, 0.09345079958438873, -0.10084061324596405, 0.005831023212522268, -0.06653685867786407, -0.046518802642822266, -0.0891256332397461, 0.02316627837717533, 0.21289560198783875, 0.03117257170379162, 0.1337406188249588, -0.05243172496557236, -0.08265405148267746, -0.0035117703955620527, 0.0280318446457386, 0.010360046289861202, 0.09838700294494629, 0.08280468732118607, -0.03951165825128555, 0.07407980412244797, 0.027351830154657364, -0.05450503155589104, 0.14522935450077057, -0.04364422336220741, -0.08752912282943726, -0.005624877288937569, -0.012682278640568256, -0.02276471257209778, 0.10845961421728134, -0.05003393813967705, 0.005424973089247942, 0.032377827912569046, 0.020417887717485428, 0.03630629554390907, -0.18249787390232086, 0.0004595220962073654, 0.021862808614969254, -0.06455869972705841, -0.025113742798566818, -0.02182350680232048, 0.052054621279239655, 0.09953272342681885, 0.005245214328169823, -0.027836913242936134, 0.009578511118888855, -0.008441010490059853, -0.06935855746269226, 0.17767325043678284, -0.11038388311862946, -0.1693577915430069, -0.1042037382721901, 0.051519960165023804, -0.0404474250972271, -0.04545212909579277, -0.004579819738864899, -0.10452314466238022, -0.06481390446424484, -0.11270593851804733, -0.023312587291002274, -0.005607659928500652, -0.007306752260774374, 0.047355640679597855, 0.011496242135763168, 0.05486425757408142, -0.13397370278835297, 0.01108731422573328, -0.032270804047584534, -0.08011535555124283, 0.015527162700891495, 0.06954526156187057, 0.06936192512512207, 0.14953184127807617, -0.01720307767391205, 0.02731860615313053, -0.021471679210662842, 0.18175673484802246, -0.09726735204458237, 0.014667303301393986, 0.10256768763065338, 0.010777078568935394, 0.04217319190502167, 0.1134273037314415, 0.03406360000371933, -0.07806511223316193, 0.02167903259396553, 0.08395767211914062, -0.02029096521437168, -0.25809526443481445, -0.06057830899953842, -0.02322925440967083, -0.06591509282588959, 0.09662239998579025, 0.050931721925735474, 0.0010202232515439391, 0.014196585863828659, -0.009891328401863575, 0.008959743194282055, -0.007850157096982002, 0.06040080636739731, 0.08251748234033585, 0.04866138845682144, 0.09285341948270798, -0.03229882940649986, -0.025968441739678383, 0.0659106969833374, -0.0014798152260482311, 0.23265202343463898, -0.05706847831606865, 0.09660864621400833, 0.014803316444158554, 0.09495055675506592, -0.010611692443490028, 0.03530658408999443, 0.024600623175501823, -0.008211432956159115, 0.014965076930820942, -0.05113111808896065, -0.008138451725244522, 0.00607337336987257, -0.011097779497504234, 0.005561324767768383, -0.06961151212453842, 0.03638111427426338, 0.014238198287785053, 0.2783264219760895, 0.04034770280122757, -0.2970677316188812, -0.05097496137022972, -0.01175632793456316, -0.030529314652085304, -0.06550867855548859, -0.0020464863628149033, 0.11819007992744446, -0.11818205565214157, 0.07067172229290009, -0.046710941940546036, 0.0878659263253212, -0.07553665339946747, -0.002502607414498925, 0.06455495953559875, 0.1365865021944046, -0.013078914023935795, 0.0664491057395935, -0.21433866024017334, 0.21350380778312683, 0.013491197489202023, 0.11678709834814072, -0.06931407004594803, 0.02568301558494568, 0.015106993727385998, 0.05127633363008499, 0.08297120034694672, -0.004451829474419355, -0.02628105692565441, -0.15129733085632324, -0.1259131282567978, 0.02930927276611328, 0.11789533495903015, -0.016804460436105728, 0.08242543786764145, -0.021668318659067154, -0.015158997848629951, 0.03676820173859596, -0.152985617518425, -0.15357327461242676, -0.12658801674842834, 0.02788134291768074, 0.03914432227611542, -0.0345473550260067, -0.05860195308923721, -0.10822207480669022, -0.023216675966978073, 0.17129074037075043, -0.01182143110781908, -0.060327813029289246, -0.14165754616260529, 0.06160151585936546, 0.14901967346668243, -0.041171107441186905, 0.014636922627687454, 0.0336611270904541, 0.1275922954082489, 0.04162311926484108, -0.07485177367925644, 0.03414439409971237, -0.059746988117694855, -0.18247437477111816, -0.05441676080226898, 0.1387583315372467, 0.07789850980043411, 0.05104202404618263, -0.01657804846763611, 0.02498379535973072, 0.007813883945345879, -0.0858364999294281, 0.0015636166790500283, 0.10024692863225937, 0.08768793940544128, 0.06714620441198349, -0.0693512111902237, 0.008711450733244419, -0.0370572954416275, -0.02383861318230629, 0.11216378957033157, 0.18020494282245636, -0.08212172240018845, 0.09752599895000458, 0.05015362426638603, -0.07735522836446762, -0.17439104616641998, 0.06168137490749359, 0.1278706043958664, 0.0262529868632555, 0.03176763281226158, -0.2047455757856369, 0.11372053623199463, 0.12729105353355408, -0.011684900149703026, 0.058386530727148056, -0.3474578261375427, -0.12836502492427826, 0.04749281331896782, 0.08163052797317505, 0.00008384983811993152, -0.10792414098978043, -0.026567190885543823, -0.02025827392935753, -0.1428883969783783, 0.13475526869297028, -0.06481514126062393, 0.10796350240707397, -0.014776203781366348, 0.09948624670505524, 0.029328133910894394, -0.04816500097513199, 0.12974117696285248, 0.05936884135007858, 0.07057301700115204, -0.04285353049635887, 0.021397655829787254, 0.04895583167672157, -0.07520755380392075, 0.051377423107624054, -0.06017889827489853, 0.06659247726202011, -0.14251108467578888, -0.02008708193898201, -0.08014815300703049, 0.04587642475962639, -0.048362795263528824, -0.04338856413960457, -0.04168276488780975, 0.047965966165065765, 0.06881176680326462, -0.04028354957699776, 0.06071818992495537, 0.003930224571377039, 0.07518289238214493, 0.07364189624786377, 0.10225548595190048, -0.03366205468773842, -0.09175081551074982, -0.005081680603325367, -0.008775665424764156, 0.04191296547651291, -0.126649409532547, 0.025245344266295433, 0.14418022334575653, 0.047047484666109085, 0.1288975179195404, 0.03252623975276947, -0.033946067094802856, -0.019724728539586067, 0.04013122245669365, -0.1217212826013565, -0.10549142956733704, 0.013912878930568695, -0.06309018284082413, -0.10996439307928085, 0.002375929383561015, 0.10325383394956589, -0.04126759245991707, -0.011866763234138489, 0.0004385439970064908, 0.027624448761343956, 0.003594352165237069, 0.2022462636232376, 0.019712120294570923, 0.06679608672857285, -0.10687784850597382, 0.1191897913813591, 0.06919308751821518, -0.09360560774803162, 0.046332016587257385, 0.11609513312578201, -0.0937708169221878, -0.009768419899046421, 0.08692702651023865, 0.1537422388792038, -0.029482271522283554, -0.03695078566670418, -0.08958928287029266, -0.10072696208953857, 0.07363535463809967, 0.11666720360517502, 0.03190808743238449, 0.00882186833769083, -0.0482909232378006, 0.015084899961948395, -0.1521710604429245, 0.06797667592763901, 0.07782431691884995, 0.060105208307504654, -0.10141940414905548, 0.14210361242294312, 0.022331958636641502, 0.029617076739668846, -0.015661589801311493, 0.01815313659608364, -0.06802820414304733, -0.010365537367761135, -0.10120640695095062, -0.006419291254132986, -0.008179694414138794, -0.0035450076684355736, -0.018751556053757668, -0.03615817427635193, -0.042636334896087646, 0.038337159901857376, -0.07158524543046951, -0.05916623771190643, -0.008092778734862804, 0.04341893270611763, -0.12921416759490967, -0.0017081426922231913, 0.002619963139295578, -0.09658548980951309, 0.07273933291435242, 0.0429028645157814, 0.005659121088683605, 0.02923688292503357, -0.12918977439403534, -0.02750796265900135, 0.03446439281105995, 0.026174357160925865, 0.0674394741654396, -0.06875275075435638, 0.0014956246595829725, -0.0215799231082201, 0.03508726879954338, 0.01785847917199135, 0.03321045637130737, -0.11250042170286179, 0.004638136830180883, -0.05824800953269005, -0.029997212812304497, -0.0708119198679924, 0.047991830855607986, 0.1120552122592926, 0.04047509282827377, 0.16714438796043396, -0.07647594064474106, 0.0393730029463768, -0.18948401510715485, -0.029802698642015457, -0.0017197742126882076, -0.03481261432170868, -0.05833088979125023, -0.02194797433912754, 0.10831467062234879, -0.053687434643507004, 0.12876944243907928, 0.007201033644378185, 0.06419443339109421, 0.038054727017879486, -0.0327860526740551, -0.0567229762673378, 0.014289799146354198, 0.1471514254808426, 0.06456049531698227, -0.024692147970199585, 0.1112382635474205, -0.014108775183558464, 0.05732501670718193, 0.05877416208386421, 0.22271816432476044, 0.13091029226779938, 0.009023969992995262, 0.07088339328765869, 0.053947582840919495, -0.13254018127918243, -0.12454593181610107, 0.1213090717792511, -0.08383583277463913, 0.1269223839044571, -0.0588560476899147, 0.19139714539051056, 0.05732734128832817, -0.1611243337392807, 0.05285812169313431, -0.03171764686703682, -0.1102282777428627, -0.1045704260468483, -0.023382216691970825, -0.08180829882621765, -0.10145144909620285, 0.029594315215945244, -0.11741197109222412, 0.058391377329826355, 0.092743419110775, 0.03155640512704849, 0.017804961651563644, 0.12609009444713593, -0.011765222065150738, 0.00015950988745316863, 0.07214418053627014, 0.02773716300725937, 0.005222411826252937, -0.0703529566526413, -0.0682787373661995, 0.033624667674303055, 0.022783182561397552, 0.08637735247612, -0.03951064869761467, -0.0013768785865977407, 0.035610321909189224, -0.010151736438274384, -0.06824013590812683, 0.03197253867983818, 0.0035509197041392326, 0.057845424860715866, 0.05451833829283714, 0.05232589691877365, 0.004871887620538473, -0.04870663583278656, 0.27154505252838135, -0.06577088683843613, -0.10290540754795074, -0.1259344220161438, 0.2130223661661148, 0.049909815192222595, -0.020944122225046158, 0.06728292256593704, -0.10127804428339005, -0.04160357639193535, 0.15662620961666107, 0.15868191421031952, -0.06522111594676971, -0.02659810148179531, -0.027266019955277443, -0.010782596655189991, -0.029109103605151176, 0.11908366531133652, 0.10271472483873367, 0.056987930089235306, -0.04741896316409111, -0.021116971969604492, -0.010088982060551643, -0.037791162729263306, -0.07101776450872421, 0.06727898120880127, 0.016153404489159584, -0.002961237682029605, -0.031923115253448486, 0.06852227449417114, -0.01491278875619173, -0.18659138679504395, 0.04389438033103943, -0.1747785061597824, -0.18608322739601135, -0.00937778688967228, 0.09136815369129181, -0.02855449728667736, 0.04179051145911217, 0.00853510107845068, -0.016035228967666626, 0.11087717860937119, -0.03228331729769707, -0.04340263456106186, -0.09563340991735458, 0.06252255290746689, -0.11005604267120361, 0.23093800246715546, 0.00005761574357165955, 0.0714399591088295, 0.09040901064872742, 0.017432164400815964, -0.12737210094928741, 0.053849007934331894, 0.04962558299303055, -0.0833410769701004, 0.028085412457585335, 0.16770976781845093, -0.04879035800695419, 0.042274486273527145, 0.0329594612121582, -0.112006276845932, -0.010070998221635818, -0.06505203247070312, -0.028967300429940224, -0.07254721969366074, -0.01957549899816513, -0.035122599452733994, 0.15002086758613586, 0.21867051720619202, -0.010481394827365875, 0.022949999198317528, -0.07141543924808502, 0.00716656306758523, 0.029446115717291832, 0.09748665243387222, -0.021917952224612236, -0.21097159385681152, 0.027197306975722313, 0.004473303910344839, 0.03423954173922539, -0.19101987779140472, -0.07340799272060394, 0.010381209664046764, -0.06438769400119781, -0.0537307932972908, 0.10920001566410065, 0.04704727232456207, 0.04190162569284439, -0.020569784566760063, -0.08844868838787079, -0.013902386650443077, 0.14448556303977966, -0.1803445965051651, -0.041904598474502563 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # albert_gpt2_summarization_xsum This model is a fine-tuned version of [](https://huggingface.co/) on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["xsum"], "model-index": [{"name": "albert_gpt2_summarization_xsum", "results": []}]}
text2text-generation
Ayham/albert_gpt2_summarization_xsum
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:xsum", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us
# albert_gpt2_summarization_xsum This model is a fine-tuned version of [](URL on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# albert_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us \n", "# albert_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 59, 35, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us \n# albert_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.08811844885349274, 0.1436471939086914, -0.0022727120667696, 0.06123565509915352, 0.1465405523777008, 0.04877171665430069, 0.10402359813451767, 0.11660027503967285, -0.08926021307706833, 0.0787343829870224, 0.08978166431188583, 0.0653696060180664, 0.06414388120174408, 0.14424915611743927, -0.03465087339282036, -0.25259071588516235, 0.017378445714712143, -0.02160538174211979, -0.0850277692079544, 0.09328197687864304, 0.08984177559614182, -0.10820842534303665, 0.07808782160282135, -0.0036631238181144, -0.16417701542377472, 0.01609715074300766, -0.044307734817266464, -0.05688004940748215, 0.09227019548416138, 0.012675810605287552, 0.09235119074583054, 0.01886563003063202, 0.12848635017871857, -0.23170295357704163, 0.003045305609703064, 0.08841431885957718, 0.03855832293629646, 0.09202378988265991, 0.06376557052135468, 0.004760076757520437, 0.13678497076034546, -0.14463189244270325, 0.10756763070821762, 0.018888456746935844, -0.07489455491304398, -0.12056038528680801, -0.08538389950990677, 0.025778280571103096, 0.08246243000030518, 0.09126272052526474, 0.004455540794879198, 0.11722736805677414, -0.09140372276306152, 0.07141787558794022, 0.17173239588737488, -0.2342359572649002, -0.058543670922517776, 0.029517877846956253, 0.05981205403804779, 0.07033316045999527, -0.10284346342086792, -0.0056730639189481735, 0.029756512492895126, 0.024242177605628967, 0.08722410351037979, -0.0012888072524219751, -0.09442498534917831, 0.013407895341515541, -0.12328310310840607, -0.013453755527734756, 0.12107314169406891, 0.03200031816959381, -0.026396382600069046, -0.10870089381933212, -0.05095523223280907, -0.11672063916921616, -0.013674497604370117, -0.032397300004959106, 0.043621376156806946, -0.03834516555070877, -0.07390769571065903, -0.040923457592725754, -0.06261112540960312, -0.060650698840618134, -0.01802847906947136, 0.14560973644256592, 0.0352407768368721, 0.012193350121378899, -0.036259762942790985, 0.09600333124399185, 0.04364543780684471, -0.11638136953115463, -0.0025745921302586794, -0.004382100887596607, -0.10761868208646774, -0.04750779643654823, -0.06160219758749008, -0.027588868513703346, 0.006750165019184351, 0.139863520860672, -0.054259371012449265, 0.09937353432178497, 0.031509142369031906, -0.011480242945253849, 0.007977772504091263, 0.1488359421491623, -0.04946557432413101, -0.04936554655432701, -0.025313017889857292, 0.07962305098772049, 0.014153214171528816, -0.024837160483002663, -0.0656759962439537, 0.000651031092274934, 0.07032210379838943, 0.054367657750844955, -0.05400789529085159, 0.023938754573464394, -0.04317086189985275, -0.028457961976528168, -0.008417596109211445, -0.12466494739055634, 0.05412157624959946, -0.0020454043988138437, -0.08382224291563034, 0.0017958383541554213, -0.0009803627617657185, 0.009892850182950497, -0.03015594184398651, 0.1271476298570633, -0.08654177933931351, -0.009814218617975712, -0.08877458423376083, -0.07370734214782715, 0.024971982464194298, -0.13278667628765106, -0.006834472995251417, -0.0389779694378376, -0.16313381493091583, -0.04704805091023445, 0.06671606004238129, -0.06015617027878761, -0.024655459448695183, -0.05226569622755051, -0.03686011955142021, 0.04182892292737961, -0.008594844490289688, 0.1673712432384491, -0.05759946256875992, 0.06070936098694801, -0.035314567387104034, 0.04062053561210632, 0.015022817999124527, 0.05026023089885712, -0.06830783188343048, 0.01087076123803854, -0.0988975465297699, 0.08500681072473526, -0.06791059672832489, 0.0047006430104374886, -0.1148199588060379, -0.0763351246714592, -0.013097560964524746, -0.0009198709740303457, 0.07579218596220016, 0.12550562620162964, -0.20043610036373138, -0.03735979646444321, 0.11125477403402328, -0.06681106239557266, -0.061460766941308975, 0.06440407782793045, -0.05248682573437691, 0.008039255626499653, 0.05027526989579201, 0.15588828921318054, 0.045572128146886826, -0.11570398509502411, -0.033197227865457535, 0.014632235281169415, 0.05738457664847374, 0.022020405158400536, 0.052403997629880905, -0.00758241955190897, 0.03071332722902298, 0.012556800618767738, -0.022975675761699677, 0.006691083312034607, -0.08062757551670074, -0.07653028517961502, -0.03676588833332062, -0.07132162898778915, -0.0026272423565387726, 0.02068202756345272, 0.02716646157205105, -0.060669053345918655, -0.10496959835290909, 0.10036129504442215, 0.12103907763957977, -0.07249278575181961, 0.019566291943192482, -0.06296204775571823, -0.018218010663986206, -0.009405306540429592, -0.01926509663462639, -0.1999732106924057, -0.1012546569108963, 0.03391536697745323, -0.07066474109888077, 0.027842959389090538, -0.012846228666603565, 0.05962309241294861, 0.05550350993871689, -0.03548374027013779, -0.012255837209522724, -0.07272699475288391, -0.0026285694912075996, -0.09573233872652054, -0.2181885689496994, -0.038458142429590225, -0.02645985037088394, 0.18633900582790375, -0.20102280378341675, -0.007192148827016354, -0.012215079739689827, 0.13937953114509583, 0.031540125608444214, -0.06919965893030167, -0.00310942018404603, 0.04589984565973282, -0.010668515227735043, -0.09721562266349792, 0.03226212039589882, 0.010503803379833698, -0.10185357928276062, -0.03856334462761879, -0.1512325257062912, 0.0004640581028070301, 0.09483234584331512, 0.06415620446205139, -0.08471762388944626, -0.03510283678770065, -0.06494836509227753, -0.03856552392244339, -0.07959597557783127, 0.022630678489804268, 0.194344162940979, 0.01532126497477293, 0.1119554191827774, -0.06157999113202095, -0.07167923450469971, 0.01180790364742279, 0.02925657108426094, -0.018296854570508003, 0.09310194104909897, 0.1258956789970398, -0.08241557329893112, 0.08171465992927551, 0.08087162673473358, -0.03375273942947388, 0.14677569270133972, -0.04009879752993584, -0.08797051757574081, -0.009515458717942238, -0.009160603396594524, -0.021346095949411392, 0.12131299078464508, -0.08639372140169144, 0.01633584499359131, 0.026033638045191765, 0.035628292709589005, 0.0477941669523716, -0.17686419188976288, 0.008327431045472622, 0.025989942252635956, -0.029437309131026268, -0.04883958771824837, -0.020542729645967484, 0.023168528452515602, 0.08930878341197968, 0.024878161028027534, 0.008996479213237762, 0.01190259400755167, -0.005708187818527222, -0.0794786736369133, 0.19249534606933594, -0.14179782569408417, -0.16599911451339722, -0.08884889632463455, 0.038388870656490326, -0.04697442054748535, -0.029640374705195427, 0.008956138975918293, -0.10891129076480865, -0.06432746350765228, -0.09420926123857498, -0.016306275501847267, -0.04993395507335663, 0.013228151947259903, 0.047807469964027405, 0.00849225465208292, 0.04386603459715843, -0.13118375837802887, 0.013185149058699608, -0.05196055397391319, -0.05844396352767944, 0.0038732015527784824, 0.07249714434146881, 0.0884307250380516, 0.12065620720386505, -0.012822280637919903, 0.024205081164836884, -0.02733832225203514, 0.20963901281356812, -0.0833682045340538, -0.009180416353046894, 0.09508413821458817, -0.012366412207484245, 0.04169098660349846, 0.11029979586601257, 0.03341019153594971, -0.09568408131599426, 0.031015025451779366, 0.07067549973726273, -0.02091890387237072, -0.25062066316604614, -0.04791080951690674, -0.03983816131949425, -0.08731481432914734, 0.107074074447155, 0.04530015587806702, -0.04819388687610626, 0.04251180216670036, -0.00427200086414814, 0.03761769458651543, -0.0258648581802845, 0.060501206666231155, 0.07267452776432037, 0.04954533651471138, 0.09801405668258667, -0.02466803416609764, -0.021333608776330948, 0.07365652173757553, -0.00023108575260266662, 0.26508980989456177, -0.02037339098751545, 0.08616221696138382, 0.0259200818836689, 0.09292175620794296, -0.01468020398169756, 0.06098853796720505, 0.02028672769665718, -0.00986769050359726, -0.010396042838692665, -0.05776076391339302, -0.03772098198533058, 0.03241956979036331, -0.007848024368286133, 0.003562238998711109, -0.09354130923748016, 0.051846519112586975, 0.026926664635539055, 0.23528292775154114, 0.021592814475297928, -0.2877485156059265, -0.05146103352308273, 0.002227945253252983, -0.025815388187766075, -0.0716961994767189, 0.002872348763048649, 0.12032383680343628, -0.10697473585605621, 0.0734262764453888, -0.06558521091938019, 0.08594711869955063, -0.05098018795251846, -0.007340497802942991, 0.060858387500047684, 0.1574907749891281, -0.010966403409838676, 0.06473590433597565, -0.2102387398481369, 0.21702629327774048, 0.025383131578564644, 0.11941950768232346, -0.08213826268911362, 0.0451592281460762, 0.012225005775690079, 0.012720629572868347, 0.08155985176563263, 0.0019149358849972486, -0.12234824895858765, -0.1393091231584549, -0.08413414657115936, 0.06666916608810425, 0.15469276905059814, -0.022584477439522743, 0.07830704748630524, -0.03901839256286621, 0.01269808504730463, 0.04994257166981697, -0.08579422533512115, -0.16705846786499023, -0.1632375866174698, 0.02869229204952717, 0.01577484980225563, -0.032743558287620544, -0.051308874040842056, -0.09742535650730133, -0.02640295773744583, 0.15499699115753174, -0.008067584596574306, -0.03903718292713165, -0.15990428626537323, 0.0781053826212883, 0.15988419950008392, -0.05349494516849518, 0.023792896419763565, 0.01782897301018238, 0.12310026586055756, 0.03206583857536316, -0.09009920060634613, 0.06134788691997528, -0.07689443975687027, -0.1539088934659958, -0.06393899768590927, 0.10145170241594315, 0.062495432794094086, 0.043463628739118576, -0.01653958112001419, 0.04246553033590317, -0.005831622984260321, -0.09687631577253342, 0.03213846683502197, 0.10418463498353958, 0.06516662985086441, 0.06481174379587173, -0.09836503863334656, 0.04612565040588379, -0.0106904786080122, -0.029408475384116173, 0.13406315445899963, 0.20051230490207672, -0.07462214678525925, 0.0999172106385231, 0.0834527388215065, -0.09173911064863205, -0.1862834095954895, 0.08124761283397675, 0.11241424828767776, 0.026004847139120102, 0.049453750252723694, -0.24248605966567993, 0.13125349581241608, 0.12493543326854706, -0.0048242779448628426, 0.05391566455364227, -0.29671812057495117, -0.12543079257011414, 0.05394873768091202, 0.10370419174432755, 0.02688831090927124, -0.11973118036985397, -0.022710425779223442, -0.04596158489584923, -0.10905203223228455, 0.14866776764392853, -0.0753406211733818, 0.11397020518779755, 0.004377377685159445, 0.06776279956102371, 0.023440171033143997, -0.042289748787879944, 0.1318417191505432, 0.028073256835341454, 0.0673619955778122, -0.030698703601956367, 0.020103033632040024, 0.0007676102686673403, -0.06056138500571251, 0.02794479951262474, -0.0928390845656395, 0.06358464807271957, -0.11282847821712494, -0.018971746787428856, -0.04459274560213089, 0.055121272802352905, -0.04429829120635986, -0.05978637561202049, -0.038168858736753464, 0.042546920478343964, 0.08208929747343063, -0.03635730221867561, 0.06895294785499573, 0.0118178129196167, 0.08001759648323059, 0.03713912516832352, 0.09972034394741058, -0.06558012962341309, -0.052817244082689285, 0.005075846333056688, -0.010251500643789768, 0.05388271063566208, -0.12526826560497284, 0.025144033133983612, 0.13532648980617523, 0.04105387255549431, 0.1350191980600357, 0.055370230227708817, -0.04241981357336044, -0.0031068711541593075, 0.05137834697961807, -0.11867808550596237, -0.12151379138231277, 0.002447146689519286, -0.03212398290634155, -0.12236439436674118, 0.01245457399636507, 0.10254491865634918, -0.033458855003118515, -0.006456519477069378, -0.017916062846779823, 0.03559647500514984, 0.005614288616925478, 0.18967069685459137, 0.006437659729272127, 0.05404544994235039, -0.10564909130334854, 0.14401812851428986, 0.0545714907348156, -0.11915573477745056, 0.07127154618501663, 0.08256565779447556, -0.083048015832901, 0.000025550541977281682, 0.06844422966241837, 0.14578081667423248, -0.027286671102046967, -0.0462186336517334, -0.08784086257219315, -0.09921132773160934, 0.05752072110772133, 0.09543142467737198, 0.02741081640124321, 0.005087649915367365, -0.04967452585697174, 0.03228865936398506, -0.15116733312606812, 0.06590890139341354, 0.05434907227754593, 0.06546636670827866, -0.10662875324487686, 0.1012972816824913, 0.02268683910369873, 0.008925773203372955, -0.01374644972383976, 0.016363056376576424, -0.09879623353481293, -0.029046058654785156, -0.08172992616891861, -0.014869466423988342, -0.031039360910654068, -0.0005938707035966218, -0.003301344346255064, -0.0329655185341835, -0.05608249828219414, 0.031125715002417564, -0.07759355753660202, -0.06034308299422264, -0.009199931286275387, 0.04466118663549423, -0.1295987218618393, 0.008873472921550274, 0.016607461497187614, -0.10333661735057831, 0.07738535851240158, 0.05451466143131256, 0.019193554297089577, 0.03680221363902092, -0.1484728902578354, -0.039288125932216644, 0.028461916372179985, 0.021355047821998596, 0.0643647164106369, -0.09335105866193771, -0.006545664742588997, -0.008751406334340572, 0.04365331679582596, 0.007487649563699961, 0.05158114433288574, -0.11257890611886978, -0.008245140314102173, -0.07325401902198792, -0.0533626526594162, -0.06323301047086716, 0.04246258735656738, 0.10664734244346619, 0.03958273306488991, 0.1597711592912674, -0.06288578361272812, 0.03406357765197754, -0.1833789348602295, -0.022605830803513527, -0.0033073658123612404, -0.03420906513929367, -0.08345905691385269, -0.02968597784638405, 0.09351962804794312, -0.05173150449991226, 0.1446366012096405, 0.00873579178005457, 0.06219922751188278, 0.03455279767513275, -0.025940829887986183, -0.045162636786699295, 0.009238678961992264, 0.20285719633102417, 0.0815180316567421, -0.017992930486798286, 0.08608072251081467, 0.006167433690279722, 0.06903109699487686, 0.0649535283446312, 0.22594054043293, 0.14949379861354828, -0.000010232721251668409, 0.08937231451272964, 0.05315669998526573, -0.13384155929088593, -0.11362022161483765, 0.15605251491069794, -0.058021217584609985, 0.12761856615543365, -0.04327503591775894, 0.18343238532543182, 0.05620991066098213, -0.1721208691596985, 0.030949532985687256, -0.04205101728439331, -0.1009170189499855, -0.13319000601768494, 0.0010146528948098421, -0.08788406103849411, -0.12100474536418915, 0.025774888694286346, -0.1314128041267395, 0.05174406245350838, 0.07980184257030487, 0.018898816779255867, 0.03504004701972008, 0.1113705262541771, -0.034986283630132675, 0.0025693331845104694, 0.05488003417849541, 0.0333305224776268, -0.001161595107987523, -0.05416664853692055, -0.09224772453308105, 0.03035084158182144, 0.008417955599725246, 0.08960764855146408, -0.05146058648824692, -0.011736179701983929, 0.05367070436477661, -0.009253728203475475, -0.06431722640991211, 0.018594970926642418, 0.0017717862501740456, 0.04023298993706703, 0.041454147547483444, 0.04431148245930672, -0.005612096283584833, -0.04575902223587036, 0.2638300359249115, -0.06756817549467087, -0.05405765771865845, -0.13603003323078156, 0.19393396377563477, 0.04246331751346588, -0.007682093419134617, 0.07049655169248581, -0.10329972207546234, -0.047607820481061935, 0.16025306284427643, 0.15587836503982544, -0.07426691055297852, -0.033287130296230316, -0.007112232968211174, -0.008891457691788673, -0.05089889466762543, 0.14380528032779694, 0.09391933679580688, 0.06560078263282776, -0.05098257586359978, -0.0202279482036829, -0.01863146759569645, -0.04812972992658615, -0.05174297094345093, 0.06532956659793854, 0.026123154908418655, -0.015499547123908997, -0.02530049905180931, 0.06880412250757217, -0.0102976830676198, -0.18206287920475006, 0.03508685901761055, -0.17696698009967804, -0.17577104270458221, -0.021448547020554543, 0.09972810745239258, -0.025477470830082893, 0.05335497856140137, 0.0007651041960343719, -0.018011610954999924, 0.10449089854955673, -0.006161874160170555, -0.05771643668413162, -0.10288091003894806, 0.07237723469734192, -0.07255221903324127, 0.2185514271259308, -0.0034251143224537373, 0.05947894603013992, 0.10034002363681793, 0.05512804538011551, -0.1269945353269577, 0.026803750544786453, 0.05902853235602379, -0.10232342034578323, 0.03743130341172218, 0.14501768350601196, -0.06644077599048615, 0.07380498945713043, 0.04772951081395149, -0.09910634905099869, 0.0037294726353138685, -0.07450316101312637, -0.03876567259430885, -0.052651502192020416, -0.01472614984959364, -0.06672430038452148, 0.16725878417491913, 0.21141070127487183, -0.014203601516783237, 0.008730041794478893, -0.07100121676921844, 0.04143504053354263, 0.03333958238363266, 0.09979412704706192, -0.029385549947619438, -0.2304425835609436, 0.018286338075995445, 0.03194455802440643, 0.01667184941470623, -0.19936364889144897, -0.07514537125825882, 0.029740039259195328, -0.05440965294837952, -0.06356669217348099, 0.09897046536207199, 0.06416957080364227, 0.024640345945954323, -0.045311182737350464, -0.12053528428077698, -0.018338408321142197, 0.1394886076450348, -0.144511878490448, -0.04208533465862274 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # albert_roberta_new_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "albert_roberta_new_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/albert_roberta_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# albert_roberta_new_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
[ "# albert_roberta_new_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# albert_roberta_new_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ 62, 42, 6, 12, 8, 3, 118, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# albert_roberta_new_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ -0.0835915058851242, 0.14600850641727448, -0.003680524881929159, 0.07549068331718445, 0.13016243278980255, 0.04182235524058342, 0.11340924352407455, 0.13130518794059753, -0.10510019958019257, 0.06539081037044525, 0.07369042932987213, 0.04622478783130646, 0.054207347333431244, 0.15337836742401123, -0.02802923694252968, -0.2326633483171463, 0.003172527998685837, -0.004277158994227648, -0.0633222684264183, 0.11498723924160004, 0.0869520977139473, -0.09776908159255981, 0.06532329320907593, -0.004654742311686277, -0.12997248768806458, 0.016546664759516716, -0.04653008654713631, -0.055256057530641556, 0.09172789752483368, -0.015866735950112343, 0.09255532175302505, 0.02722439356148243, 0.14551186561584473, -0.2085278332233429, 0.0027039754204452038, 0.08864591270685196, 0.03588063642382622, 0.09139858186244965, 0.06720911711454391, -0.014875159598886967, 0.12573480606079102, -0.166298970580101, 0.10157632827758789, 0.014183145016431808, -0.0808786004781723, -0.09754015505313873, -0.10147316753864288, 0.057402629405260086, 0.09552594274282455, 0.11453437805175781, 0.00720625277608633, 0.12514354288578033, -0.10146960616111755, 0.08096739649772644, 0.17439110577106476, -0.2108173817396164, -0.05437683314085007, 0.07497621327638626, 0.0297103151679039, 0.0593552440404892, -0.09674612432718277, -0.019367199391126633, 0.024278374388813972, 0.027247291058301926, 0.10778872668743134, -0.0015265209367498755, -0.0968995988368988, 0.00284576159901917, -0.12771311402320862, -0.030239176005125046, 0.09493891894817352, 0.02382843755185604, -0.007148949895054102, -0.10049638897180557, -0.053518395870923996, -0.15040527284145355, -0.025079915300011635, -0.017299776896834373, 0.03731810301542282, -0.05460308492183685, -0.05793248116970062, -0.048091061413288116, -0.06903419643640518, -0.054590269923210144, -0.01728176698088646, 0.10086068511009216, 0.041862085461616516, 0.008831637911498547, -0.04954248294234276, 0.12283068895339966, 0.06062370166182518, -0.13172931969165802, 0.003107671858742833, 0.0138118090108037, -0.0926903560757637, -0.055939774960279465, -0.022525012493133545, -0.0298896674066782, -0.008941245265305042, 0.13354390859603882, -0.06604617089033127, 0.08037330955266953, 0.01775864139199257, -0.012950888834893703, -0.015010254457592964, 0.140996053814888, -0.04400267452001572, -0.049341682344675064, -0.0139072360470891, 0.11435973644256592, 0.0016306100878864527, -0.010287645272910595, -0.06536496430635452, -0.00375690427608788, 0.075318843126297, 0.05779413506388664, -0.041701361536979675, 0.04405848681926727, -0.01924816519021988, -0.022032639011740685, 0.035762615501880646, -0.1389542818069458, 0.04256531223654747, 0.01567988097667694, -0.10068552196025848, -0.00813582818955183, 0.017782678827643394, -0.005485666450113058, -0.045593954622745514, 0.12735697627067566, -0.077571801841259, 0.006229655351489782, -0.08154906332492828, -0.06676937639713287, 0.02106092870235443, -0.10128041356801987, -0.04304714500904083, -0.029374506324529648, -0.19737660884857178, -0.06257116049528122, 0.02444525994360447, -0.05882197245955467, -0.03413402661681175, -0.06395042687654495, -0.07279527932405472, 0.019110029563307762, -0.004846071358770132, 0.16610795259475708, -0.0659044161438942, 0.06849745661020279, -0.02471785619854927, 0.027056200429797173, 0.07050827890634537, 0.04718394577503204, -0.06637027859687805, -0.0004648812173400074, -0.10307563096284866, 0.08881153911352158, -0.08116006851196289, 0.004106951877474785, -0.10787257552146912, -0.09466798603534698, 0.008978217840194702, -0.01289463136345148, 0.04717054218053818, 0.14266552031040192, -0.19308781623840332, -0.030867701396346092, 0.13226325809955597, -0.05818679928779602, -0.011521832086145878, 0.05611031875014305, -0.061160530894994736, 0.0021885433234274387, 0.06301072984933853, 0.12368863821029663, 0.0779372900724411, -0.12304423749446869, -0.011466077528893948, -0.006822274997830391, 0.03478033095598221, 0.009383889846503735, 0.022970454767346382, 0.010478389449417591, 0.05894634127616882, 0.008875660598278046, -0.050904013216495514, 0.004078196361660957, -0.09294185042381287, -0.0687636137008667, -0.037757713347673416, -0.0704045295715332, 0.020280763506889343, 0.02350895293056965, 0.03941749036312103, -0.053396452218294144, -0.10664764791727066, 0.0857907086610794, 0.1270419955253601, -0.05261624604463577, 0.007419066037982702, -0.07239188998937607, -0.03586727753281593, 0.006553075276315212, -0.024572480469942093, -0.19550645351409912, -0.11069337278604507, 0.029948465526103973, -0.05635475367307663, 0.043181102722883224, -0.0038111377507448196, 0.07008691877126694, 0.04153457656502724, -0.040010370314121246, -0.020310968160629272, -0.06714481115341187, -0.0034859469160437584, -0.08728574216365814, -0.19527721405029297, -0.031986843794584274, -0.021757131442427635, 0.15956604480743408, -0.21843551099300385, 0.01947730965912342, -0.00796029157936573, 0.1517610102891922, 0.02871658280491829, -0.056469712406396866, 0.023819157853722572, 0.049422457814216614, 0.008939575403928757, -0.09466857463121414, 0.027698619291186333, -0.009946310892701149, -0.07129380851984024, -0.024502625688910484, -0.1365288347005844, -0.019368914887309074, 0.07230637222528458, 0.09620659798383713, -0.10622067749500275, 0.0095086470246315, -0.07127542048692703, -0.0507558174431324, -0.09706863760948181, 0.026071548461914062, 0.2019544392824173, 0.03800400346517563, 0.1275564283132553, -0.0468585379421711, -0.08143892884254456, -0.005537917837500572, 0.026990989223122597, 0.006503826938569546, 0.09458617120981216, 0.07729007303714752, -0.047227151691913605, 0.07475567609071732, 0.012514608912169933, -0.04537038877606392, 0.14105410873889923, -0.04689266160130501, -0.08940458297729492, -0.007572915870696306, -0.01590990461409092, -0.02337116189301014, 0.10014469176530838, -0.03745974972844124, -0.0019082307117059827, 0.03145374357700348, 0.026946358382701874, 0.034477490931749344, -0.1788148432970047, -0.0027625467628240585, 0.014254797250032425, -0.062402643263339996, -0.016711978241801262, -0.012868566438555717, 0.05322078987956047, 0.10081282258033752, 0.003946592565625906, -0.03464875370264053, 0.011284422129392624, -0.010702566243708134, -0.07500775903463364, 0.17358151078224182, -0.10735145211219788, -0.16636614501476288, -0.10150133073329926, 0.031147850677371025, -0.03546005114912987, -0.04720776900649071, 0.0015646233223378658, -0.10014797747135162, -0.06368974596261978, -0.11401405930519104, -0.020366981625556946, 0.005243164021521807, -0.005871300585567951, 0.038427047431468964, 0.015665093436837196, 0.047620560973882675, -0.13542810082435608, 0.01310715265572071, -0.03806978091597557, -0.0769488662481308, 0.01987689547240734, 0.06347495317459106, 0.06510105729103088, 0.14345279335975647, -0.015089890919625759, 0.030614711344242096, -0.02009361982345581, 0.1876552253961563, -0.0922791063785553, 0.01696157641708851, 0.10214222967624664, 0.012134172953665257, 0.04768960550427437, 0.11595848947763443, 0.03643817827105522, -0.07294556498527527, 0.027184896171092987, 0.0892365574836731, -0.019283488392829895, -0.2608503997325897, -0.05905436724424362, -0.01929626800119877, -0.06698489189147949, 0.09325341135263443, 0.06080883368849754, 0.0014580104034394026, 0.011289980262517929, -0.008677303791046143, -0.008997290395200253, -0.009042123332619667, 0.06199140474200249, 0.0808999091386795, 0.04949605464935303, 0.08638351410627365, -0.03184210881590843, -0.03411588445305824, 0.06344614923000336, 0.0008195509435608983, 0.22588825225830078, -0.0516243614256382, 0.07924860715866089, 0.014464443549513817, 0.0930875763297081, -0.014491145499050617, 0.0361623652279377, 0.024964144453406334, -0.0034455833956599236, 0.011916855350136757, -0.05104876682162285, -0.004841928370296955, 0.00984716136008501, -0.018742306157946587, 0.0008108585025183856, -0.06153859570622444, 0.033704135566949844, 0.014763145707547665, 0.2897891104221344, 0.03663468360900879, -0.29141807556152344, -0.04617797210812569, -0.01617484726011753, -0.0340808667242527, -0.053696855902671814, -0.009906361810863018, 0.10042412579059601, -0.12176330387592316, 0.06676444411277771, -0.051877427846193314, 0.08560065180063248, -0.0821598693728447, -0.003522018902003765, 0.06120208650827408, 0.12960992753505707, -0.017583992332220078, 0.06559442728757858, -0.20944638550281525, 0.21901977062225342, 0.012979866936802864, 0.11989991366863251, -0.07171820104122162, 0.022598309442400932, 0.020919406786561012, 0.04050140827894211, 0.07704382389783859, -0.009192770346999168, -0.036869458854198456, -0.14635001122951508, -0.13208886981010437, 0.023436585441231728, 0.1123894453048706, -0.013869248330593109, 0.08251870423555374, -0.016955425962805748, -0.016229543834924698, 0.036579959094524384, -0.1579420566558838, -0.15114443004131317, -0.11584033817052841, 0.038253262639045715, 0.03450765088200569, -0.03527659550309181, -0.054682765156030655, -0.11753390729427338, -0.021937336772680283, 0.16482675075531006, 0.011417675763368607, -0.05521952360868454, -0.14483746886253357, 0.05843522399663925, 0.15056024491786957, -0.043085064738988876, 0.010209429077804089, 0.039091531187295914, 0.12436604499816895, 0.03961065784096718, -0.07881831377744675, 0.02907191403210163, -0.05485183745622635, -0.18339070677757263, -0.05659886822104454, 0.13553597033023834, 0.08999846130609512, 0.04607120156288147, -0.01967591978609562, 0.02889210730791092, 0.006921985652297735, -0.08044032007455826, 0.0027857415843755007, 0.08795994520187378, 0.08772425353527069, 0.061077944934368134, -0.07148782908916473, -0.0016241425182670355, -0.0456278994679451, -0.017931245267391205, 0.10339730232954025, 0.17984651029109955, -0.07652685791254044, 0.10255542397499084, 0.05268013849854469, -0.07175566256046295, -0.169738307595253, 0.05797227472066879, 0.12496258318424225, 0.019557179883122444, 0.04133165627717972, -0.20192226767539978, 0.11105295270681381, 0.11995366960763931, -0.015804406255483627, 0.0505184680223465, -0.34823453426361084, -0.1201895922422409, 0.049630049616098404, 0.08025436848402023, 0.005420127883553505, -0.10998085886240005, -0.02817712351679802, -0.0019587913993746042, -0.1394142210483551, 0.13529668748378754, -0.06120283529162407, 0.10916916280984879, -0.01695176586508751, 0.11346038430929184, 0.025970209389925003, -0.045964211225509644, 0.11458002775907516, 0.0743066817522049, 0.07248944044113159, -0.04745902493596077, 0.006682340521365404, 0.049570780247449875, -0.07621847838163376, 0.054268497973680496, -0.05157748609781265, 0.06819713860750198, -0.14729410409927368, -0.015456180088222027, -0.07968635112047195, 0.04647861421108246, -0.047762006521224976, -0.04155148193240166, -0.03812913969159126, 0.0548970066010952, 0.07005228102207184, -0.04153444245457649, 0.06999202072620392, 0.00950195174664259, 0.08265139907598495, 0.08463852107524872, 0.08745972812175751, -0.024221545085310936, -0.09454218298196793, -0.013531009666621685, -0.0068903896026313305, 0.03184535726904869, -0.1252458691596985, 0.01837289333343506, 0.13829416036605835, 0.04788150638341904, 0.13223819434642792, 0.029851913452148438, -0.042804304510354996, -0.019746525213122368, 0.03971116989850998, -0.12785010039806366, -0.0922931581735611, 0.021606141701340675, -0.05880191922187805, -0.11788515746593475, 0.005895542446523905, 0.10131167620420456, -0.03940393403172493, -0.019050875678658485, 0.005061148665845394, 0.030818868428468704, 0.008306645788252354, 0.21004857122898102, 0.022850485518574715, 0.06474293768405914, -0.1063099056482315, 0.1223282665014267, 0.07020673900842667, -0.08769364655017853, 0.04293544963002205, 0.12792855501174927, -0.09668925404548645, -0.011882721446454525, 0.09932681918144226, 0.1632591336965561, -0.029676897451281548, -0.02920832671225071, -0.09257262945175171, -0.08938770741224289, 0.07440944015979767, 0.12962684035301208, 0.032187387347221375, 0.012895957566797733, -0.05141856148838997, 0.01206387672573328, -0.14210990071296692, 0.069332554936409, 0.09169173240661621, 0.058262038975954056, -0.08734972029924393, 0.13279104232788086, 0.018382349982857704, 0.03583409637212753, -0.016609443351626396, 0.011742478236556053, -0.069438636302948, -0.014016525819897652, -0.07784341275691986, -0.0035709466319531202, -0.01598416082561016, -0.0038759829476475716, -0.02131994627416134, -0.04557371139526367, -0.04034340754151344, 0.03585905581712723, -0.07194023579359055, -0.05584787204861641, -0.009654194116592407, 0.048860400915145874, -0.13193745911121368, -0.0009449376375414431, 0.005634127650409937, -0.09462080895900726, 0.07095342874526978, 0.04771234467625618, -0.00120004010386765, 0.026462901383638382, -0.1267281472682953, -0.026890691369771957, 0.03155560418963432, 0.015628254041075706, 0.06910642236471176, -0.07572191208600998, 0.0010181089164689183, -0.02685883268713951, 0.031843382865190506, 0.018364379182457924, 0.039808809757232666, -0.11669351160526276, 0.012893627397716045, -0.07075124233961105, -0.029924431815743446, -0.06887573003768921, 0.0390234999358654, 0.11820748448371887, 0.036869924515485764, 0.1679116040468216, -0.07777754217386246, 0.045118045061826706, -0.1907925307750702, -0.02757740393280983, 0.0004137087380513549, -0.041248537600040436, -0.06373867392539978, -0.01989091746509075, 0.10662936419248581, -0.05154269561171532, 0.12229842692613602, 0.013799437321722507, 0.060473959892988205, 0.03753020241856575, -0.03558500111103058, -0.06397794187068939, 0.009984511882066727, 0.1370479017496109, 0.059519506990909576, -0.02441979944705963, 0.11535464972257614, -0.017601318657398224, 0.045560888946056366, 0.04026264697313309, 0.21980923414230347, 0.12995970249176025, 0.01796327531337738, 0.07006770372390747, 0.06446178257465363, -0.129081591963768, -0.11401639878749847, 0.11145828664302826, -0.07137902081012726, 0.11721284687519073, -0.06027914211153984, 0.1935296356678009, 0.06120564788579941, -0.15380626916885376, 0.05588584020733833, -0.02699139527976513, -0.10301536321640015, -0.09989330917596817, -0.019831834360957146, -0.08111370354890823, -0.10318923741579056, 0.03410162031650543, -0.11415330320596695, 0.056068215519189835, 0.08972200006246567, 0.02960069663822651, 0.01728850044310093, 0.11596435308456421, -0.013607065193355083, 0.0018652405124157667, 0.07033126056194305, 0.024480333551764488, 0.005289916414767504, -0.07191327959299088, -0.06471654772758484, 0.024070287123322487, 0.0323672778904438, 0.08608966320753098, -0.03836210444569588, 0.01209052000194788, 0.03709575906395912, -0.02011672593653202, -0.07442730665206909, 0.036209408193826675, 0.0029427935369312763, 0.06122243404388428, 0.05983762443065643, 0.05695381388068199, 0.0020281276665627956, -0.04506710171699524, 0.27342233061790466, -0.061198893934488297, -0.09857470542192459, -0.1268150955438614, 0.20907717943191528, 0.05151872709393501, -0.02425994724035263, 0.059914480894804, -0.10765887051820755, -0.039294321089982986, 0.15552417933940887, 0.16208253800868988, -0.049026649445295334, -0.02404893934726715, -0.028842682018876076, -0.01034644152969122, -0.01885385252535343, 0.1038508415222168, 0.09424998611211777, 0.04938221350312233, -0.048592522740364075, -0.022299788892269135, -0.014313412830233574, -0.03935650363564491, -0.06913184374570847, 0.06369122117757797, 0.018157469108700752, -0.0032143632415682077, -0.03257375210523605, 0.07783634960651398, -0.023011837154626846, -0.19872692227363586, 0.0402652770280838, -0.17340782284736633, -0.18812845647335052, -0.005010951776057482, 0.08795858919620514, -0.02270723693072796, 0.047121480107307434, 0.010145297273993492, -0.017117170616984367, 0.10325781255960464, -0.031443506479263306, -0.04059261456131935, -0.10254829376935959, 0.06323111057281494, -0.11285057663917542, 0.2229098677635193, -0.0011463647242635489, 0.06899931281805038, 0.08732148259878159, 0.017833290621638298, -0.12893164157867432, 0.05569813773036003, 0.04600713029503822, -0.08053280413150787, 0.02567358873784542, 0.1695559173822403, -0.04616929218173027, 0.04199574515223503, 0.03468361869454384, -0.11193600296974182, -0.010153552517294884, -0.07136763632297516, -0.021922757849097252, -0.07404327392578125, -0.01533499825745821, -0.024342963472008705, 0.15418581664562225, 0.21460700035095215, -0.008521223440766335, 0.025934427976608276, -0.07290726155042648, 0.0027735282201319933, 0.027402441948652267, 0.07846518605947495, -0.0273479912430048, -0.21644525229930878, 0.026873856782913208, 0.0067500765435397625, 0.03426632285118103, -0.1880962997674942, -0.08036839216947556, 0.01502945739775896, -0.06657559424638748, -0.04852748662233353, 0.10881142318248749, 0.044205401092767715, 0.03898797184228897, -0.02012675814330578, -0.08445126563310623, -0.012543385848402977, 0.14746007323265076, -0.1924433559179306, -0.04091915488243103 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert_bert_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "bert_bert_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/bert_bert_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# bert_bert_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
[ "# bert_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# bert_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ 62, 39, 6, 12, 8, 3, 118, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# bert_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ -0.0836322233080864, 0.13003499805927277, -0.0036609431263059378, 0.06656044721603394, 0.14319048821926117, 0.035764411091804504, 0.10391192883253098, 0.10630470514297485, -0.11642326414585114, 0.05639564245939255, 0.0744226723909378, 0.07282853126525879, 0.04188397154211998, 0.1614290177822113, -0.037874989211559296, -0.20966392755508423, -0.00046634895261377096, -0.0037944393698126078, -0.07353300601243973, 0.11625850200653076, 0.08163006603717804, -0.11089460551738739, 0.05839577317237854, -0.0043879724107682705, -0.1410665214061737, 0.02017403393983841, -0.038426484912633896, -0.05219809338450432, 0.09571494907140732, -0.0066842325031757355, 0.09808970242738724, 0.035412952303886414, 0.14842043817043304, -0.19469931721687317, 0.0047026039101183414, 0.1006668210029602, 0.04422600567340851, 0.09517709910869598, 0.059061743319034576, -0.00665148114785552, 0.11893510818481445, -0.15035711228847504, 0.10401181876659393, 0.0153048662468791, -0.07803073525428772, -0.10360532253980637, -0.1031852588057518, 0.052019596099853516, 0.0885004997253418, 0.1212303414940834, 0.006317269988358021, 0.13178043067455292, -0.11467162519693375, 0.08034009486436844, 0.17969204485416412, -0.20378758013248444, -0.04750535637140274, 0.05579020082950592, 0.019667811691761017, 0.06203006953001022, -0.09095834195613861, -0.018213938921689987, 0.027523206546902657, 0.02407236024737358, 0.0994597002863884, 0.006427351851016283, -0.09872211515903473, -0.0012656101025640965, -0.12833289802074432, -0.01934397779405117, 0.05877910554409027, 0.0228269025683403, -0.013347324915230274, -0.08728263527154922, -0.05360642820596695, -0.14203664660453796, -0.02916254661977291, -0.010667303577065468, 0.03817019611597061, -0.05847849324345589, -0.07390718907117844, -0.03931594640016556, -0.06444390118122101, -0.049362339079380035, -0.03220241889357567, 0.12975041568279266, 0.045795783400535583, 0.006473247427493334, -0.04777613654732704, 0.11250968277454376, 0.07456440478563309, -0.12087073922157288, 0.02413209341466427, 0.01494398433715105, -0.08478407561779022, -0.04971366003155708, -0.03227754682302475, -0.01376373041421175, -0.009330234490334988, 0.12386522442102432, -0.0798560380935669, 0.07736983150243759, 0.01974179781973362, -0.014772817492485046, -0.01533682830631733, 0.140629842877388, -0.050641484558582306, -0.02953118085861206, -0.010395112447440624, 0.11417099088430405, 0.0034039481543004513, -0.015547139570116997, -0.06676670908927917, 0.013676434755325317, 0.06677019596099854, 0.05736280605196953, -0.04293772950768471, 0.046530645340681076, -0.019844183698296547, -0.013436184264719486, 0.04612482339143753, -0.1359165459871292, 0.04438052698969841, 0.018816610798239708, -0.09961442649364471, -0.009152359329164028, 0.005811622366309166, -0.00018945205374620855, -0.050005532801151276, 0.1394709199666977, -0.07689134776592255, 0.005449860356748104, -0.08463999629020691, -0.08349131047725677, 0.011378084309399128, -0.11145433038473129, -0.039480920881032944, -0.03681232035160065, -0.16341201961040497, -0.05990453064441681, 0.02251850627362728, -0.05341489613056183, -0.04132958874106407, -0.05827442184090614, -0.07663760334253311, 0.02106018364429474, -0.01304522156715393, 0.19255641102790833, -0.06987885385751724, 0.06926403194665909, -0.03269873559474945, 0.027140283957123756, 0.0568685382604599, 0.04632860794663429, -0.06375198066234589, -0.01566263474524021, -0.10065058618783951, 0.08433254808187485, -0.08149050176143646, 0.0038855329621583223, -0.10517149418592453, -0.09428833425045013, 0.030275341123342514, -0.0065831588581204414, 0.0656597912311554, 0.12684491276741028, -0.20056141912937164, -0.03497956320643425, 0.1064119040966034, -0.047546014189720154, -0.0028230769094079733, 0.05054517835378647, -0.057459671050310135, -0.012896359898149967, 0.06531120836734772, 0.127569317817688, 0.05475877597928047, -0.12871362268924713, -0.004031335469335318, 0.00025749296764843166, 0.038122307509183884, 0.009237015619874, 0.012720141559839249, 0.007388587109744549, 0.05412551760673523, 0.005296480841934681, -0.03729385510087013, 0.008337334729731083, -0.10395634174346924, -0.06246165931224823, -0.031144317239522934, -0.07869409769773483, 0.01482783816754818, 0.02600840851664543, 0.03324580565094948, -0.06145821884274483, -0.11273588985204697, 0.08500771969556808, 0.11929971724748611, -0.05378688871860504, 0.020435767248272896, -0.07822183519601822, -0.03237240016460419, 0.01377220917493105, -0.020678045228123665, -0.20332802832126617, -0.10803817212581635, 0.02200060710310936, -0.051136910915374756, 0.04278869181871414, -0.013837489299476147, 0.05828468129038811, 0.0426594503223896, -0.03436683490872383, -0.014857010915875435, -0.06149103119969368, 0.003799211699515581, -0.09864837676286697, -0.21540753543376923, -0.03092101775109768, -0.0157085619866848, 0.1535687893629074, -0.2154506891965866, 0.011161742731928825, -0.019004935398697853, 0.14262215793132782, 0.02283654361963272, -0.05794810131192207, 0.007358049973845482, 0.06334496289491653, 0.01097320020198822, -0.08799896389245987, 0.03817698732018471, -0.0034736471716314554, -0.07093466073274612, -0.01390411239117384, -0.13794578611850739, -0.024410896003246307, 0.08629438281059265, 0.07641462236642838, -0.10592375695705414, 0.026594890281558037, -0.07345511764287949, -0.03581526130437851, -0.09266549348831177, 0.04336096718907356, 0.21913407742977142, 0.034909382462501526, 0.13463178277015686, -0.05017237737774849, -0.08325840532779694, -0.00528173241764307, 0.024429501965641975, 0.016423771157860756, 0.11009040474891663, 0.08711729198694229, -0.03613831847906113, 0.06971199810504913, 0.019201042130589485, -0.04079621657729149, 0.12128232419490814, -0.042851850390434265, -0.0850750133395195, 0.010532929562032223, -0.03649192303419113, -0.026809662580490112, 0.09262831509113312, -0.034241460263729095, 0.007095058914273977, 0.029973842203617096, 0.028649812564253807, 0.03265433758497238, -0.19232642650604248, 0.0021113029215484858, 0.014971571043133736, -0.05216400325298309, -0.05162271112203598, -0.023303093388676643, 0.05954783037304878, 0.10889516025781631, 0.00886219460517168, -0.03440316393971443, 0.004756920505315065, -0.005544943735003471, -0.06994607299566269, 0.1726401001214981, -0.11604733020067215, -0.16349537670612335, -0.0700058564543724, 0.003356419038027525, -0.025693561881780624, -0.05280996486544609, 0.0069192806258797646, -0.12395027279853821, -0.07136933505535126, -0.11651962995529175, -0.023884087800979614, 0.014534251764416695, 0.0003704531700350344, 0.0563078336417675, 0.010662198998034, 0.047525737434625626, -0.13741394877433777, 0.013628975488245487, -0.06085289269685745, -0.07860399037599564, 0.019445622339844704, 0.08815465122461319, 0.05969158187508583, 0.14989447593688965, -0.021362824365496635, 0.026456128805875778, -0.019740022718906403, 0.2056449055671692, -0.0884881392121315, 0.007086444180458784, 0.10168307274580002, 0.0117768170312047, 0.043477848172187805, 0.12135633081197739, 0.043387047946453094, -0.07886315882205963, 0.030688825994729996, 0.09679379314184189, -0.01945364847779274, -0.2577648460865021, -0.06680383533239365, -0.014038742519915104, -0.09202840924263, 0.08466372638940811, 0.05173623561859131, -0.028723930940032005, 0.005045519210398197, -0.0035035221371799707, -0.012424595654010773, -0.010138175450265408, 0.048642758280038834, 0.07442054152488708, 0.05379648879170418, 0.09515612572431564, -0.029590576887130737, -0.0255903173238039, 0.06502983719110489, -0.014620769768953323, 0.21849781274795532, -0.07072614133358002, 0.058419372886419296, 0.026821447536349297, 0.08809758722782135, -0.007993625476956367, 0.031377919018268585, 0.016466617584228516, -0.016062375158071518, 0.011140052229166031, -0.04593691602349281, -0.00008017875370569527, -0.011686708778142929, -0.01575431413948536, -0.005321688484400511, -0.07193505018949509, 0.031716253608465195, 0.010880865156650543, 0.2783588469028473, 0.03602082282304764, -0.28322866559028625, -0.04699520021677017, -0.028857577592134476, -0.032260023057460785, -0.046155575662851334, -0.00739523908123374, 0.09043677896261215, -0.10821990668773651, 0.06180013716220856, -0.03164760768413544, 0.08865083754062653, -0.07067475467920303, 0.004308858420699835, 0.06373672187328339, 0.14641806483268738, -0.019085632637143135, 0.05333150550723076, -0.2156587690114975, 0.22216781973838806, 0.0063123032450675964, 0.12555648386478424, -0.06258819252252579, 0.013577586971223354, 0.014312823303043842, 0.03654959425330162, 0.055860333144664764, -0.004400334786623716, -0.008712515234947205, -0.16570958495140076, -0.11931318044662476, 0.028951579704880714, 0.11881618201732635, 0.00709824962541461, 0.0800052061676979, -0.020737331360578537, -0.014764583669602871, 0.03963004797697067, -0.16143330931663513, -0.15858551859855652, -0.12117581069469452, 0.025844920426607132, 0.03431689739227295, -0.040732335299253464, -0.05445754528045654, -0.12385919690132141, -0.040427226573228836, 0.17065995931625366, 0.03276887163519859, -0.055838845670223236, -0.14247837662696838, 0.0658901184797287, 0.1436123549938202, -0.03945916146039963, 0.009716653265058994, 0.03557157516479492, 0.11679082363843918, 0.047692567110061646, -0.08470144122838974, 0.03503463417291641, -0.06181197240948677, -0.18783831596374512, -0.05616698041558266, 0.12357430160045624, 0.09072104096412659, 0.0521463043987751, -0.023169653490185738, 0.026491612195968628, 0.0030192213598638773, -0.0784330889582634, 0.0038551161997020245, 0.08334264904260635, 0.08520577847957611, 0.06035090982913971, -0.07047634571790695, -0.010094298049807549, -0.04832195118069649, -0.024222172796726227, 0.1076856404542923, 0.20205077528953552, -0.07649004459381104, 0.10670755803585052, 0.07698160409927368, -0.06572481244802475, -0.18507450819015503, 0.0615827850997448, 0.13283151388168335, 0.0066438778303563595, 0.022678017616271973, -0.19947776198387146, 0.12737679481506348, 0.11809952557086945, -0.008613628335297108, 0.04396767169237137, -0.3392573297023773, -0.1280973255634308, 0.05489421263337135, 0.09357478469610214, 0.031336747109889984, -0.0978403389453888, -0.01360680814832449, -0.01650889404118061, -0.1395394206047058, 0.1470368355512619, -0.05174997076392174, 0.10716898739337921, -0.0021989932283759117, 0.10207151621580124, 0.022071242332458496, -0.04314017295837402, 0.09917093068361282, 0.056538067758083344, 0.06835244596004486, -0.043899890035390854, 0.009406263940036297, 0.026732446625828743, -0.06094750761985779, 0.02528472989797592, -0.037958044558763504, 0.07206585258245468, -0.11097042262554169, -0.012381224893033504, -0.08544202893972397, 0.03373872488737106, -0.044379882514476776, -0.04333409294486046, -0.03657694533467293, 0.048413973301649094, 0.07703159004449844, -0.04791468009352684, 0.05935148522257805, 0.000445941201178357, 0.08554740250110626, 0.05517522990703583, 0.09141524136066437, -0.07313628494739532, -0.08178930729627609, -0.003268037224188447, -0.01471923291683197, 0.033559270203113556, -0.1077243760228157, 0.022159604355692863, 0.1347871720790863, 0.05142151564359665, 0.11454510688781738, 0.0403989814221859, -0.03231838718056679, -0.018649529665708542, 0.03464636206626892, -0.12342305481433868, -0.07805132120847702, 0.03454853966832161, -0.048078056424856186, -0.10314866900444031, 0.013538896106183529, 0.10242463648319244, -0.03220250830054283, -0.023341339081525803, 0.0009708137949928641, 0.025631796568632126, 0.015647167339920998, 0.21387235820293427, 0.028045648708939552, 0.06316198408603668, -0.12256786227226257, 0.12429116666316986, 0.06506655365228653, -0.09640040993690491, 0.04801357164978981, 0.12820999324321747, -0.1050366684794426, -0.008151991292834282, 0.11016669869422913, 0.14650875329971313, -0.02774093672633171, -0.027813801541924477, -0.08738865703344345, -0.09949783235788345, 0.08753271400928497, 0.14620152115821838, 0.030128546059131622, -0.0026638652198016644, -0.055578283965587616, 0.015910983085632324, -0.15268197655677795, 0.06010264903306961, 0.0819905623793602, 0.05687440186738968, -0.07670658081769943, 0.14113974571228027, 0.0291079543530941, 0.03516209125518799, -0.018815258517861366, 0.014652570709586143, -0.07761635631322861, -0.008533444255590439, -0.08048874139785767, -0.010561365634202957, -0.004254166502505541, -0.003547973232343793, -0.024495944380760193, -0.04954613000154495, -0.04159789904952049, 0.04186602309346199, -0.07745427638292313, -0.05592607334256172, 0.001184896333143115, 0.03657225891947746, -0.1377299278974533, 0.007461724802851677, 0.0011525247246026993, -0.09229755401611328, 0.06805773079395294, 0.04274293780326843, -0.008258810266852379, 0.03466998413205147, -0.16493096947669983, -0.05355677381157875, 0.030452869832515717, 0.018494918942451477, 0.08647861331701279, -0.07288919389247894, -0.0016106506809592247, -0.017251236364245415, 0.04299238324165344, 0.013139443472027779, 0.061331551522016525, -0.11063770204782486, 0.0130571024492383, -0.061828549951314926, -0.034151848405599594, -0.07093232125043869, 0.03937313333153725, 0.11094976961612701, 0.04425175115466118, 0.17486847937107086, -0.07917547225952148, 0.033552661538124084, -0.174617737531662, -0.03322957828640938, -0.006219165399670601, -0.04580198600888252, -0.07441836595535278, -0.013474980369210243, 0.11588436365127563, -0.051190998405218124, 0.1222291961312294, 0.013648152351379395, 0.06442104279994965, 0.04117147997021675, -0.033121224492788315, -0.06875713914632797, 0.014215957373380661, 0.1466483771800995, 0.06349639594554901, -0.02040272019803524, 0.112082839012146, -0.008090537041425705, 0.05088561028242111, 0.055833104997873306, 0.23382961750030518, 0.13307921588420868, 0.04043798893690109, 0.06465014070272446, 0.04427741467952728, -0.13995467126369476, -0.11071529239416122, 0.10706670582294464, -0.10390080511569977, 0.12663331627845764, -0.07006154954433441, 0.21785560250282288, 0.044207438826560974, -0.14177002012729645, 0.052953507751226425, -0.03256405144929886, -0.09814442694187164, -0.10566068440675735, 0.0021425196900963783, -0.07688412070274353, -0.10615747421979904, 0.038219645619392395, -0.11328044533729553, 0.06837619096040726, 0.10160215198993683, 0.031095828860998154, 0.03278937190771103, 0.10282395780086517, -0.00981744471937418, 0.003811189206317067, 0.06069153547286987, 0.02265147864818573, -0.005137431900948286, -0.06855245679616928, -0.06816327571868896, 0.03031904250383377, 0.027935229241847992, 0.0860789567232132, -0.032845817506313324, -0.013661738485097885, 0.04783695191144943, -0.013968897983431816, -0.07359766215085983, 0.043163176625967026, -0.0034722720738500357, 0.07436695694923401, 0.06193891540169716, 0.05224153771996498, 0.012441671453416348, -0.04768248647451401, 0.2875998914241791, -0.06035176292061806, -0.11763030290603638, -0.11865680664777756, 0.24540172517299652, 0.058423519134521484, -0.02270139381289482, 0.06524676829576492, -0.1012369841337204, -0.05523664131760597, 0.1706286370754242, 0.16948539018630981, -0.0699123740196228, -0.02687234804034233, -0.03048834763467312, -0.011636354960501194, -0.03241271153092384, 0.14072687923908234, 0.12156219780445099, 0.054046742618083954, -0.05243859812617302, -0.02697274275124073, -0.024452965706586838, -0.035062745213508606, -0.07069062441587448, 0.0670444518327713, 0.025487560778856277, -0.007563526276499033, -0.040669336915016174, 0.0657624900341034, -0.010052373632788658, -0.21190671622753143, 0.025214090943336487, -0.16774187982082367, -0.18522432446479797, -0.004830953665077686, 0.10056556761264801, -0.027865687385201454, 0.046788815408945084, 0.01791277341544628, -0.0074010491371154785, 0.0944540873169899, -0.034803636372089386, -0.01842552050948143, -0.10180363059043884, 0.0641721859574318, -0.1336800754070282, 0.2265782505273819, -0.006440763361752033, 0.07402003556489944, 0.09270337224006653, 0.031497783958911896, -0.12385857850313187, 0.06925220787525177, 0.0414753295481205, -0.10505245625972748, 0.030666273087263107, 0.1564202904701233, -0.04376975819468498, 0.03724005073308945, 0.03687085583806038, -0.12279953807592392, 0.003610367653891444, -0.09714224934577942, -0.03315246105194092, -0.056883394718170166, -0.01943281851708889, -0.031620658934116364, 0.14586688578128815, 0.2112511545419693, -0.006457324139773846, 0.03091377019882202, -0.08112921565771103, -0.010205326601862907, 0.01786203309893608, 0.08074552565813065, -0.030413733795285225, -0.23216935992240906, 0.03581393510103226, 0.01562643237411976, 0.03595555201172829, -0.18798446655273438, -0.06314834952354431, 0.001857021707110107, -0.069252610206604, -0.05616999417543411, 0.105789415538311, 0.041346751153469086, 0.04006423428654671, -0.024909615516662598, -0.08435888588428497, -0.010858593508601189, 0.15135128796100616, -0.1868637055158615, -0.039003886282444 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert_distilgpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "bert_distilgpt2_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/bert_distilgpt2_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# bert_distilgpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
[ "# bert_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# bert_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ 62, 43, 6, 12, 8, 3, 118, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# bert_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ -0.08565270900726318, 0.14421911537647247, -0.003633857239037752, 0.06637522578239441, 0.13230305910110474, 0.043946411460638046, 0.10684926807880402, 0.12582048773765564, -0.1026172786951065, 0.05636240169405937, 0.07957235723733902, 0.0471426360309124, 0.053762517869472504, 0.15361301600933075, -0.03361210227012634, -0.2249380350112915, 0.005231246817857027, -0.0032376921735703945, -0.07437819242477417, 0.12237050384283066, 0.08672430366277695, -0.0965644046664238, 0.057549890130758286, -0.008838996291160583, -0.1325339525938034, 0.024251319468021393, -0.043762750923633575, -0.05822024494409561, 0.09327242523431778, -0.008783859200775623, 0.09451138228178024, 0.038000620901584625, 0.14806976914405823, -0.20620977878570557, 0.0013383064651861787, 0.09013412147760391, 0.03953682258725166, 0.08981127291917801, 0.06354527175426483, -0.014636130072176456, 0.11750546097755432, -0.1674613207578659, 0.09901705384254456, 0.01400040928274393, -0.07494492828845978, -0.11549006402492523, -0.09955954551696777, 0.0760919451713562, 0.09546862542629242, 0.11532050371170044, 0.003735968843102455, 0.12235016375780106, -0.11362840235233307, 0.08033505082130432, 0.17518402636051178, -0.20645944774150848, -0.05009489506483078, 0.0580187626183033, 0.016525112092494965, 0.057917576283216476, -0.08741947263479233, -0.016628915444016457, 0.027755260467529297, 0.02201230078935623, 0.09887555241584778, 0.00552733801305294, -0.09437195956707001, 0.002855701604858041, -0.12540635466575623, -0.0379413478076458, 0.10581182688474655, 0.02304934151470661, -0.012710279785096645, -0.10698778927326202, -0.046418506652116776, -0.14010459184646606, -0.013609548099339008, -0.008892429061233997, 0.03511253744363785, -0.06569656729698181, -0.05836297199130058, -0.05265131592750549, -0.07567740976810455, -0.05028391629457474, -0.013493684120476246, 0.1039605364203453, 0.04572663828730583, 0.011874931864440441, -0.041455209255218506, 0.12309329211711884, 0.06925971060991287, -0.12504401803016663, 0.0030548616778105497, 0.012960226275026798, -0.09141791611909866, -0.049484241753816605, -0.022599203512072563, -0.01608247123658657, -0.013293251395225525, 0.127419576048851, -0.07068881392478943, 0.07427985966205597, 0.021531105041503906, -0.012222867459058762, -0.021602675318717957, 0.14390693604946136, -0.04269013553857803, -0.04055441915988922, -0.019608577713370323, 0.11647727340459824, 0.002925651613622904, -0.008087166585028172, -0.06879465281963348, -0.00028565595857799053, 0.0728011354804039, 0.0627647340297699, -0.03976541385054588, 0.04583324119448662, -0.020006094127893448, -0.022043591365218163, 0.03937021642923355, -0.14708657562732697, 0.037829864770174026, 0.019561326131224632, -0.09654060751199722, -0.001967945136129856, 0.02228507213294506, -0.011616888456046581, -0.05877991020679474, 0.11935234069824219, -0.07342026382684708, 0.007744940463453531, -0.07914960384368896, -0.07153326272964478, 0.021634746342897415, -0.10777605324983597, -0.0416565015912056, -0.03296579048037529, -0.19140486419200897, -0.06076540797948837, 0.02411539852619171, -0.0556885227560997, -0.049384262412786484, -0.058078113943338394, -0.07987187057733536, 0.01771385967731476, -0.010232384316623211, 0.17727652192115784, -0.06113940477371216, 0.07129628211259842, -0.022897738963365555, 0.02413034252822399, 0.07521722465753555, 0.048219915479421616, -0.06463876366615295, 0.0006374411168508232, -0.09800957143306732, 0.09272852540016174, -0.08606785535812378, -0.003632793901488185, -0.10706369578838348, -0.10406982153654099, 0.023488974198698997, -0.011110354214906693, 0.043921951204538345, 0.1465739607810974, -0.1811991035938263, -0.03722880408167839, 0.12878404557704926, -0.061352215707302094, -0.000410668202675879, 0.061588022857904434, -0.05153335630893707, -0.01585865393280983, 0.059850338846445084, 0.12988492846488953, 0.0757877305150032, -0.12557515501976013, -0.013031018897891045, 0.007743676193058491, 0.04434150084853172, 0.006268024910241365, 0.032754085958004, 0.008355120196938515, 0.05831535905599594, 0.009154798462986946, -0.06762268394231796, 0.003985207993537188, -0.0951479896903038, -0.06808438152074814, -0.035899050533771515, -0.07649686187505722, 0.04077064245939255, 0.030858848243951797, 0.03846452385187149, -0.05581505224108696, -0.11694318056106567, 0.09286770969629288, 0.12437704205513, -0.059788208454847336, 0.005561159458011389, -0.07904288172721863, -0.03676670789718628, 0.009981751441955566, -0.019875485450029373, -0.1962047815322876, -0.12437481433153152, 0.02016330510377884, -0.05525757744908333, 0.041049376130104065, -0.005210323724895716, 0.06565158069133759, 0.038357485085725784, -0.04531437158584595, -0.025226783007383347, -0.07264871895313263, -0.002376934979110956, -0.08427207916975021, -0.1925252079963684, -0.04958036169409752, -0.018894821405410767, 0.1771961897611618, -0.22999830543994904, 0.024807296693325043, -0.007794674951583147, 0.14596286416053772, 0.02295963652431965, -0.0610431544482708, 0.02223607897758484, 0.05422656610608101, 0.00986489001661539, -0.08964664489030838, 0.028961816802620888, -0.01148670632392168, -0.07388981431722641, -0.011666424572467804, -0.13180842995643616, -0.023033102974295616, 0.0710396096110344, 0.10240129381418228, -0.1099926307797432, 0.015328075736761093, -0.07631773501634598, -0.046245068311691284, -0.09776618331670761, 0.01987612433731556, 0.20179253816604614, 0.03896552696824074, 0.13218864798545837, -0.047183431684970856, -0.08862056583166122, -0.004886606242507696, 0.017053229734301567, 0.00022224332496989518, 0.1097114309668541, 0.06926862895488739, -0.04767907038331032, 0.08248487114906311, 0.017425313591957092, -0.04898348078131676, 0.1410147249698639, -0.05318231135606766, -0.09354661405086517, 0.0011532075004652143, -0.016721483319997787, -0.025983363389968872, 0.10242817550897598, -0.0461825355887413, -0.0076855760999023914, 0.03655797243118286, 0.01911947876214981, 0.03548930585384369, -0.18198469281196594, -0.004407666623592377, 0.017941011115908623, -0.06318559497594833, -0.038358014076948166, -0.02784150280058384, 0.060596201568841934, 0.10294022411108017, -0.0004064778913743794, -0.03239419311285019, 0.019727816805243492, -0.008918076753616333, -0.07884453237056732, 0.17632488906383514, -0.10810188949108124, -0.18042340874671936, -0.10312650352716446, 0.04218931496143341, -0.03347030282020569, -0.054631903767585754, 0.00010881950584007427, -0.11104399710893631, -0.06388908624649048, -0.11312925815582275, -0.020069511607289314, 0.0008546951576136053, -0.011091852560639381, 0.05068749189376831, 0.01601729728281498, 0.05101519078016281, -0.12758319079875946, 0.013799651525914669, -0.03867073729634285, -0.08136492967605591, 0.026548760011792183, 0.06325829029083252, 0.05849882587790489, 0.14702285826206207, -0.019104283303022385, 0.03242822363972664, -0.02059970796108246, 0.17953261733055115, -0.09660208225250244, 0.010758462361991405, 0.11308520287275314, 0.013089711777865887, 0.04322121664881706, 0.11104149371385574, 0.03417683765292168, -0.06323020905256271, 0.01868077553808689, 0.09102008491754532, -0.026754310354590416, -0.2664133608341217, -0.06632552295923233, -0.014242913573980331, -0.06782405078411102, 0.09076103568077087, 0.057834457606077194, -0.01159743219614029, 0.009754836559295654, -0.014250918291509151, -0.018344003707170486, 0.0003032059466931969, 0.05310431867837906, 0.08730225265026093, 0.04326449707150459, 0.08934110403060913, -0.03342752903699875, -0.028407879173755646, 0.0627160295844078, 0.0018799942918121815, 0.22974756360054016, -0.06313309818506241, 0.07717975229024887, 0.013341397047042847, 0.09751265496015549, -0.012157591059803963, 0.03589213639497757, 0.02984464541077614, -0.007249397225677967, 0.01908036880195141, -0.04575789347290993, -0.004602842032909393, 0.0039429618045687675, -0.004704279359430075, -0.006356362719088793, -0.0687033012509346, 0.0325445793569088, 0.006205516401678324, 0.29988208413124084, 0.04608331248164177, -0.2903895676136017, -0.056018318980932236, -0.02041909471154213, -0.0406937375664711, -0.049827344715595245, -0.011339954100549221, 0.10834972560405731, -0.1285858303308487, 0.07162158191204071, -0.043058574199676514, 0.08241827040910721, -0.07228399068117142, -0.002114724600687623, 0.056110627949237823, 0.11516512930393219, -0.01113719679415226, 0.07010066509246826, -0.20684807002544403, 0.23013558983802795, 0.00588268181309104, 0.11399934440851212, -0.07180333882570267, 0.022570544853806496, 0.013642659410834312, 0.04376579448580742, 0.08537961542606354, -0.005802174098789692, -0.015521548688411713, -0.15781205892562866, -0.1320246160030365, 0.022615235298871994, 0.11294081062078476, -0.016481930390000343, 0.08448156714439392, -0.01818097196519375, -0.018291030079126358, 0.02877132222056389, -0.1527784764766693, -0.14051327109336853, -0.12621864676475525, 0.035903897136449814, 0.028643375262618065, -0.03292666748166084, -0.05927500128746033, -0.12143820524215698, -0.023716187104582787, 0.1897154152393341, 0.0031768777407705784, -0.0690932422876358, -0.14652331173419952, 0.06514447927474976, 0.14638391137123108, -0.040631551295518875, 0.0024868769105523825, 0.04210539534687996, 0.12525209784507751, 0.04064323753118515, -0.08744649589061737, 0.02039826661348343, -0.061923548579216, -0.19149455428123474, -0.04409605264663696, 0.14162154495716095, 0.0840001255273819, 0.055128034204244614, -0.015419586561620235, 0.023930981755256653, 0.00958779826760292, -0.07425820082426071, 0.009312647394835949, 0.08740045875310898, 0.08767416328191757, 0.07077690213918686, -0.0719989761710167, -0.010007526725530624, -0.04685629904270172, -0.022005440667271614, 0.10828274488449097, 0.17483405768871307, -0.08203065395355225, 0.10732433199882507, 0.062849260866642, -0.06921476125717163, -0.18051019310951233, 0.04550808668136597, 0.13895291090011597, 0.019523439928889275, 0.033030830323696136, -0.2086341381072998, 0.11667175590991974, 0.12824074923992157, -0.012218459509313107, 0.03655731678009033, -0.3448140323162079, -0.1191825196146965, 0.06143776327371597, 0.07287361472845078, 0.01291893981397152, -0.10857968777418137, -0.02795429155230522, 0.001774353557266295, -0.1328325718641281, 0.14307476580142975, -0.05103881657123566, 0.10418666154146194, -0.008718091063201427, 0.1159777119755745, 0.02976842224597931, -0.04875772073864937, 0.11843141168355942, 0.07211878895759583, 0.059087470173835754, -0.04430581256747246, 0.011012084782123566, 0.050374794751405716, -0.07312929630279541, 0.06284931302070618, -0.044453155249357224, 0.06510284543037415, -0.1483229100704193, -0.018169691786170006, -0.09094946831464767, 0.043850719928741455, -0.043245118111371994, -0.035930220037698746, -0.028832586482167244, 0.0538531169295311, 0.06618054211139679, -0.044293362647295, 0.0682515799999237, 0.001616787281818688, 0.08895234763622284, 0.08734940737485886, 0.09983354061841965, -0.03281202167272568, -0.10901188105344772, -0.020534100010991096, -0.013440541923046112, 0.035478077828884125, -0.10599680244922638, 0.01817920058965683, 0.13859395682811737, 0.05868976190686226, 0.13236097991466522, 0.029016900807619095, -0.040097326040267944, -0.01990995928645134, 0.03660861402750015, -0.12567047774791718, -0.1057906299829483, 0.012005914933979511, -0.057226814329624176, -0.12001251429319382, 0.009675750508904457, 0.10397806763648987, -0.03568031266331673, -0.024169085547327995, -0.0006710365414619446, 0.034399792551994324, 0.012016857974231243, 0.21419492363929749, 0.025300389155745506, 0.07642609626054764, -0.1140097826719284, 0.11622480303049088, 0.06839253753423691, -0.10572738200426102, 0.04895821213722229, 0.1296127736568451, -0.09626027196645737, -0.01650610752403736, 0.09260515123605728, 0.12834231555461884, -0.03409643471240997, -0.035589247941970825, -0.09101080149412155, -0.10411173850297928, 0.08061942458152771, 0.12437707185745239, 0.03588134050369263, 0.011546198278665543, -0.04665988311171532, 0.005675940308719873, -0.14450721442699432, 0.06896640360355377, 0.08996263146400452, 0.053570058196783066, -0.09067915380001068, 0.14810791611671448, 0.02156723663210869, 0.048981551080942154, -0.01625949703156948, 0.012702461332082748, -0.06664244830608368, -0.006915847770869732, -0.07967670261859894, -0.004852179437875748, -0.008887602016329765, -0.004407401196658611, -0.01637725904583931, -0.04361788183450699, -0.03289366513490677, 0.04341063275933266, -0.08050739765167236, -0.05671292915940285, -0.005192540120333433, 0.04589569941163063, -0.13602586090564728, 0.0007163126138038933, 0.004264301620423794, -0.1008656769990921, 0.0725121945142746, 0.05276897922158241, -0.004989020060747862, 0.030941564589738846, -0.1203724592924118, -0.03906925767660141, 0.03040863387286663, 0.022873753681778908, 0.06955427676439285, -0.07116647809743881, 0.003620614530518651, -0.018455389887094498, 0.026051515713334084, 0.025186877697706223, 0.04181154817342758, -0.11250297725200653, 0.009026167914271355, -0.06374956667423248, -0.035498492419719696, -0.07713450491428375, 0.049971453845500946, 0.10924378037452698, 0.04351082816720009, 0.1677437573671341, -0.08164792507886887, 0.04218740016222, -0.18354550004005432, -0.030770858749747276, -0.002962027909234166, -0.03489572927355766, -0.05445418506860733, -0.012255504727363586, 0.11077066510915756, -0.04522331804037094, 0.11279413104057312, 0.009817810729146004, 0.06209871545433998, 0.0396268256008625, -0.034678421914577484, -0.05653361976146698, 0.013864539563655853, 0.13595910370349884, 0.0563349649310112, -0.02224501222372055, 0.11178108304738998, -0.023934664204716682, 0.03451007604598999, 0.04886704683303833, 0.22507469356060028, 0.12991651892662048, 0.010085038840770721, 0.06827714294195175, 0.0584050789475441, -0.13831107318401337, -0.1237284392118454, 0.09756697714328766, -0.08544742316007614, 0.12930095195770264, -0.06534697115421295, 0.21110008656978607, 0.05843507871031761, -0.15394680202007294, 0.05887867137789726, -0.027738993987441063, -0.10642275959253311, -0.1033710390329361, -0.027512602508068085, -0.07694574445486069, -0.0972733274102211, 0.036943159997463226, -0.11709874123334885, 0.062352631241083145, 0.10411021113395691, 0.0331268385052681, 0.023949116468429565, 0.11289256066083908, -0.009041170589625835, -0.004917129408568144, 0.07286128401756287, 0.022210724651813507, 0.0012958016013726592, -0.07850474864244461, -0.05730035528540611, 0.039875589311122894, 0.028093673288822174, 0.0862545445561409, -0.03533022478222847, 0.0032550578471273184, 0.033830676227808, -0.01640435680747032, -0.0741298645734787, 0.03424007073044777, 0.0001431675918865949, 0.07313849031925201, 0.06406274437904358, 0.051875580102205276, 0.002803168026730418, -0.05396213382482529, 0.2745093107223511, -0.06568722426891327, -0.10609681904315948, -0.13688135147094727, 0.2357300966978073, 0.04569629952311516, -0.026293795555830002, 0.07185395061969757, -0.10563766956329346, -0.03947703167796135, 0.15423886477947235, 0.16514989733695984, -0.037126194685697556, -0.0219395961612463, -0.033256758004426956, -0.00989510864019394, -0.019016452133655548, 0.11462366580963135, 0.10196945816278458, 0.07354522496461868, -0.05433012172579765, -0.019336652010679245, -0.011042380705475807, -0.03673002868890762, -0.07962755858898163, 0.07621848583221436, 0.020305441692471504, -0.005904259625822306, -0.0379609689116478, 0.0756181925535202, -0.01026306115090847, -0.2003004401922226, 0.04352821037173271, -0.174668088555336, -0.18557710945606232, -0.009910224005579948, 0.08735424280166626, -0.026344574987888336, 0.04677141457796097, 0.014596418477594852, -0.016923658549785614, 0.10980558395385742, -0.029906144365668297, -0.048632267862558365, -0.09837284684181213, 0.0734042152762413, -0.11480998992919922, 0.22653785347938538, -0.0020935824140906334, 0.07887451350688934, 0.08609149605035782, 0.01220808643847704, -0.1287512630224228, 0.05540221557021141, 0.042484525591135025, -0.0823279321193695, 0.028263777494430542, 0.1777513176202774, -0.04434856027364731, 0.04300583526492119, 0.02923089824616909, -0.12546685338020325, -0.015043459832668304, -0.07184994965791702, -0.027024144306778908, -0.06265051662921906, -0.01170350331813097, -0.025666726753115654, 0.14554497599601746, 0.22200816869735718, -0.011599359102547169, 0.02325889654457569, -0.07837458699941635, -0.0037614961620420218, 0.029186850413680077, 0.07891841232776642, -0.01626935973763466, -0.21429657936096191, 0.02700238674879074, -0.02088344842195511, 0.03808528557419777, -0.1955210417509079, -0.07437499612569809, 0.007582777179777622, -0.062163095921278, -0.051375336945056915, 0.10313259810209274, 0.0541902631521225, 0.041079360991716385, -0.018788928166031837, -0.06442293524742126, -0.011234506033360958, 0.14507612586021423, -0.19087538123130798, -0.03296332061290741 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert_gpt2_summarization_cnndm This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "bert_gpt2_summarization_cnndm", "results": []}]}
text2text-generation
Ayham/bert_gpt2_summarization_cnndm
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# bert_gpt2_summarization_cnndm This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# bert_gpt2_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# bert_gpt2_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 62, 39, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# bert_gpt2_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.08415017277002335, 0.13109640777111053, -0.0036026390735059977, 0.06847946345806122, 0.14904868602752686, 0.0400402657687664, 0.09218212962150574, 0.10759356617927551, -0.12021993845701218, 0.0549393929541111, 0.07555412501096725, 0.07317793369293213, 0.04446369409561157, 0.1527496725320816, -0.035794034600257874, -0.21816176176071167, 0.001969436649233103, -0.005707705859094858, -0.08389947563409805, 0.1168837919831276, 0.08395232260227203, -0.10945018380880356, 0.05915781855583191, -0.0022822485771030188, -0.1464368849992752, 0.01629970595240593, -0.038477808237075806, -0.050089430063962936, 0.09748312085866928, -0.008581137284636497, 0.10111545771360397, 0.033617328852415085, 0.1500013917684555, -0.1976446509361267, 0.003707048948854208, 0.09647930413484573, 0.04792650416493416, 0.0966041088104248, 0.054992083460092545, 0.0016020426992326975, 0.11803976446390152, -0.14774608612060547, 0.09913893789052963, 0.014694571495056152, -0.07725159823894501, -0.10602500289678574, -0.0963372066617012, 0.04387005418539047, 0.0874728411436081, 0.117630735039711, 0.007322568446397781, 0.1254408061504364, -0.11390656232833862, 0.08053460717201233, 0.17138521373271942, -0.21326224505901337, -0.0524037703871727, 0.06462114304304123, 0.02760288491845131, 0.05776713788509369, -0.09473781287670135, -0.015167200937867165, 0.026710517704486847, 0.022850172594189644, 0.09581650048494339, 0.004509255290031433, -0.10376659035682678, -0.000401794706704095, -0.12723203003406525, -0.018107928335666656, 0.07128843665122986, 0.025215359404683113, -0.01205376349389553, -0.08744510263204575, -0.05605299770832062, -0.14016807079315186, -0.025366883724927902, -0.019560186192393303, 0.03858156129717827, -0.0565374381840229, -0.06940589100122452, -0.04265844076871872, -0.06776104867458344, -0.05684976279735565, -0.028690382838249207, 0.128693625330925, 0.041396159678697586, 0.007379996590316296, -0.0431356281042099, 0.12079216539859772, 0.06377752870321274, -0.11641846597194672, 0.01946214772760868, 0.012193076312541962, -0.09316112101078033, -0.053553808480501175, -0.034652773290872574, -0.01213890127837658, -0.007181091699749231, 0.12399891018867493, -0.07397519797086716, 0.07800080627202988, 0.0238826721906662, -0.014428935945034027, -0.014116933569312096, 0.14000748097896576, -0.059054892510175705, -0.030211197212338448, -0.011431768536567688, 0.10834009200334549, -0.0007546358974650502, -0.01428352203220129, -0.06786082684993744, -0.0014133332297205925, 0.0714825913310051, 0.057884279638528824, -0.04748885706067085, 0.04771367833018303, -0.028462056070566177, -0.017509041354060173, 0.03845096006989479, -0.13214953243732452, 0.0429757758975029, 0.012713571079075336, -0.10072272270917892, -0.011414076201617718, 0.010123182088136673, 0.005107426550239325, -0.04371802508831024, 0.135623961687088, -0.07787592709064484, 0.00594743387773633, -0.0860556811094284, -0.07634203881025314, 0.008780390955507755, -0.10714111477136612, -0.034950267523527145, -0.035361018031835556, -0.17820672690868378, -0.05578336492180824, 0.03053315356373787, -0.05664701759815216, -0.043307460844516754, -0.05812045559287071, -0.07470265030860901, 0.02080673910677433, -0.013574831187725067, 0.19730399549007416, -0.06792663037776947, 0.0706726536154747, -0.029232146218419075, 0.02829880081117153, 0.04648732393980026, 0.04262426495552063, -0.06870211660861969, -0.012440339662134647, -0.11040416359901428, 0.0829399824142456, -0.07668668031692505, 0.005737473256886005, -0.10550105571746826, -0.09660010039806366, 0.020208872854709625, -0.012479772791266441, 0.06648646295070648, 0.12777438759803772, -0.20163999497890472, -0.028577687218785286, 0.10629047453403473, -0.0462787039577961, -0.00958053395152092, 0.05204334855079651, -0.05508778244256973, 0.00012760602112393826, 0.06127864122390747, 0.1290290355682373, 0.059931688010692596, -0.12789149582386017, -0.002347890054807067, 0.003513473318889737, 0.031092070043087006, 0.012574059888720512, 0.018325075507164, 0.008237753063440323, 0.048156339675188065, 0.006632015574723482, -0.0379469096660614, 0.007379624526947737, -0.10302653908729553, -0.0664605125784874, -0.03349511697888374, -0.07716045528650284, 0.01603560522198677, 0.026667824015021324, 0.03190428763628006, -0.05758015438914299, -0.11196202784776688, 0.08485201746225357, 0.12510357797145844, -0.05560389161109924, 0.020145786926150322, -0.07551524043083191, -0.030987534672021866, 0.010217586532235146, -0.023790854960680008, -0.2041635364294052, -0.11088044941425323, 0.02038811892271042, -0.05055251345038414, 0.04034361615777016, -0.00915064662694931, 0.05950651690363884, 0.044924359768629074, -0.03655211627483368, -0.013501827605068684, -0.06329117715358734, 0.0017159025883302093, -0.09858844429254532, -0.21360896527767181, -0.0313393734395504, -0.014126181602478027, 0.1702263355255127, -0.2198447287082672, 0.008817754685878754, -0.01973254233598709, 0.14597420394420624, 0.018608245998620987, -0.058418646454811096, 0.008349070325493813, 0.055636148899793625, 0.007671731989830732, -0.08953329175710678, 0.03852652385830879, -0.000594373035710305, -0.07775048166513443, -0.026941074058413506, -0.13370828330516815, -0.025368520990014076, 0.084151990711689, 0.06501369178295135, -0.10599955171346664, 0.017962994053959846, -0.0707819014787674, -0.041397880762815475, -0.08595354110002518, 0.04034855216741562, 0.21637658774852753, 0.03164545074105263, 0.1328735649585724, -0.051522284746170044, -0.08115949481725693, -0.003299191128462553, 0.026622559875249863, 0.018261820077896118, 0.10341739654541016, 0.09687838703393936, -0.04278038069605827, 0.06816674768924713, 0.02670997381210327, -0.041269220411777496, 0.1278379112482071, -0.03764108568429947, -0.08329865336418152, 0.004466789774596691, -0.02861378714442253, -0.029493406414985657, 0.09815119951963425, -0.04721754789352417, 0.013727767392992973, 0.030428212136030197, 0.02779722958803177, 0.035963740199804306, -0.19150352478027344, -0.0007971828454174101, 0.015548721887171268, -0.05226865038275719, -0.04626326262950897, -0.022716060280799866, 0.05666336417198181, 0.10738768428564072, 0.010689043439924717, -0.03585699200630188, 0.006937995087355375, -0.005536368116736412, -0.06663873046636581, 0.17707891762256622, -0.11680597811937332, -0.17038625478744507, -0.07223966717720032, 0.02032279595732689, -0.028147097676992416, -0.05103719234466553, -0.001966065028682351, -0.11797938495874405, -0.06783340871334076, -0.10919792950153351, -0.02682659775018692, 0.0047345892526209354, -0.002676847390830517, 0.05673176050186157, 0.011504124850034714, 0.05203067511320114, -0.13479484617710114, 0.016498377546668053, -0.05802981182932854, -0.0847320705652237, 0.020278755575418472, 0.0891818106174469, 0.06226874142885208, 0.15076926350593567, -0.023312289267778397, 0.025482594966888428, -0.020672189071774483, 0.20384643971920013, -0.09105268120765686, 0.007238788530230522, 0.10026552528142929, 0.01214522309601307, 0.040899716317653656, 0.1241474375128746, 0.040330905467271805, -0.08448798209428787, 0.031220605596899986, 0.08904743194580078, -0.015988560393452644, -0.2504841983318329, -0.0663122609257698, -0.016689959913492203, -0.08786828815937042, 0.08571179211139679, 0.04690294340252876, -0.02700044959783554, 0.011356620118021965, -0.0011809052666649222, -0.0038891404401510954, -0.009253067895770073, 0.051206137984991074, 0.07717733085155487, 0.0532783642411232, 0.09780998528003693, -0.024756580591201782, -0.028755763545632362, 0.064568430185318, -0.008725195191800594, 0.22600938379764557, -0.06715535372495651, 0.06765224784612656, 0.023562489077448845, 0.10005638748407364, -0.010431500151753426, 0.03503543138504028, 0.021352354437112808, -0.015137992799282074, 0.009302906692028046, -0.04656407982110977, -0.005777055863291025, -0.010224135592579842, -0.016260487958788872, -0.0005845544510520995, -0.07338181138038635, 0.0386183075606823, 0.01264659408479929, 0.26677900552749634, 0.0393754281103611, -0.28549447655677795, -0.049608685076236725, -0.02733754925429821, -0.030216913670301437, -0.055440519005060196, -0.004957667086273432, 0.0948527455329895, -0.11476671695709229, 0.06248438358306885, -0.04123735800385475, 0.0856865644454956, -0.06374513357877731, 0.002466015052050352, 0.06871698796749115, 0.14828264713287354, -0.017116064205765724, 0.05995643138885498, -0.2126028686761856, 0.2189406007528305, 0.007204887457191944, 0.1232517883181572, -0.06265652924776077, 0.017452780157327652, 0.012107678689062595, 0.03952503576874733, 0.06365419179201126, -0.004613781347870827, -0.011343270540237427, -0.16212691366672516, -0.11633964627981186, 0.03447597101330757, 0.12254171073436737, 0.0001886269310489297, 0.0792645514011383, -0.02525380440056324, -0.010693991556763649, 0.04245045408606529, -0.15737737715244293, -0.16601230204105377, -0.11847112327814102, 0.02671916037797928, 0.03395902365446091, -0.036911848932504654, -0.05234239250421524, -0.11756846308708191, -0.039426837116479874, 0.1773647516965866, 0.023881766945123672, -0.05275142565369606, -0.14046289026737213, 0.0743577852845192, 0.15220162272453308, -0.038153793662786484, 0.014023832976818085, 0.030010588467121124, 0.11431743949651718, 0.05341733619570732, -0.08101603388786316, 0.03962438926100731, -0.06293804943561554, -0.18504679203033447, -0.05592045560479164, 0.12441317737102509, 0.07997483760118484, 0.051630135625600815, -0.02049359865486622, 0.02749064192175865, 0.00012650668213609606, -0.08201270550489426, 0.001296202652156353, 0.07594606280326843, 0.08762648701667786, 0.06976454704999924, -0.07193364202976227, -0.001459381659515202, -0.041937366127967834, -0.02757342904806137, 0.11467141658067703, 0.2043263018131256, -0.07821866869926453, 0.0999758169054985, 0.07407963275909424, -0.06931285560131073, -0.18340949714183807, 0.06615511327981949, 0.13223451375961304, 0.011028039269149303, 0.027373850345611572, -0.20341573655605316, 0.1326262205839157, 0.12350207567214966, -0.009304563514888287, 0.048512525856494904, -0.3335120379924774, -0.130886971950531, 0.0509650893509388, 0.09459929913282394, 0.02623656392097473, -0.09584231674671173, -0.012107953429222107, -0.01990964636206627, -0.14272725582122803, 0.1486431360244751, -0.05435620993375778, 0.10993500798940659, -0.0030037816613912582, 0.10570208728313446, 0.02530176006257534, -0.044985171407461166, 0.10575174540281296, 0.051653023809194565, 0.07027977705001831, -0.04372597485780716, 0.011551634408533573, 0.029177077114582062, -0.059022825211286545, 0.02834869548678398, -0.03352959081530571, 0.07144765555858612, -0.11157186329364777, -0.01085048820823431, -0.08675164729356766, 0.03887087106704712, -0.045130785554647446, -0.048567578196525574, -0.0384974367916584, 0.05276038497686386, 0.07190670073032379, -0.04323660954833031, 0.0552755631506443, 0.004067361820489168, 0.0792861208319664, 0.05473864823579788, 0.09812426567077637, -0.06852457672357559, -0.06930068135261536, -0.0007563260733149946, -0.01006710808724165, 0.035031288862228394, -0.10475067049264908, 0.02572960965335369, 0.14092978835105896, 0.04898790642619133, 0.11702712625265121, 0.04295647144317627, -0.028762372210621834, -0.01860900968313217, 0.0385560616850853, -0.1276623010635376, -0.0821378156542778, 0.028445539996027946, -0.05924012511968613, -0.10354232788085938, 0.01705970987677574, 0.10452420264482498, -0.03443159535527229, -0.018042700365185738, -0.0002983833837788552, 0.024174226447939873, 0.008518273010849953, 0.20968732237815857, 0.025245821103453636, 0.06220482289791107, -0.11913994699716568, 0.12559814751148224, 0.06595500558614731, -0.08863317221403122, 0.04444687068462372, 0.11415056884288788, -0.10474476963281631, -0.006924960296601057, 0.09786047041416168, 0.1427859216928482, -0.028833571821451187, -0.028189850971102715, -0.08725189417600632, -0.0936623141169548, 0.07894163578748703, 0.1323065161705017, 0.030601266771554947, -0.0016739676939323545, -0.052496228367090225, 0.020452024415135384, -0.16049274802207947, 0.06105419248342514, 0.07772223651409149, 0.05812530964612961, -0.09121154993772507, 0.1488647311925888, 0.025942372158169746, 0.03449550271034241, -0.01979321427643299, 0.014896686188876629, -0.07357398420572281, -0.008814699947834015, -0.0764315128326416, -0.016058893874287605, -0.006284288130700588, -0.005010535474866629, -0.018943114206194878, -0.04694005101919174, -0.044369135051965714, 0.0393436960875988, -0.07678388804197311, -0.05292518809437752, 0.0030202760826796293, 0.033444810658693314, -0.13598111271858215, 0.0018101002788171172, 0.00210851663723588, -0.088550865650177, 0.07203052937984467, 0.044699378311634064, 0.0019084709929302335, 0.03636537492275238, -0.16651467978954315, -0.04895472899079323, 0.032202597707509995, 0.02198738604784012, 0.08144479990005493, -0.07257421314716339, -0.0033135011326521635, -0.016839228570461273, 0.04867158457636833, 0.013025525026023388, 0.05587303638458252, -0.11162832379341125, 0.004445159807801247, -0.06258147954940796, -0.03570982441306114, -0.06963969767093658, 0.038908205926418304, 0.10340055078268051, 0.04506896063685417, 0.17602162063121796, -0.07412685453891754, 0.03264234587550163, -0.18076330423355103, -0.03269294649362564, -0.007120358757674694, -0.04080776870250702, -0.07460946589708328, -0.014712510630488396, 0.11373729258775711, -0.05433664470911026, 0.12302705645561218, 0.009571859613060951, 0.06898786872625351, 0.03701159730553627, -0.031959958374500275, -0.06563244014978409, 0.014843511395156384, 0.1504199504852295, 0.07237005978822708, -0.019562572240829468, 0.11542516201734543, -0.001434283098205924, 0.05457520857453346, 0.05942368879914284, 0.23102465271949768, 0.1279008835554123, 0.029400015249848366, 0.06456305831670761, 0.044504664838314056, -0.1365111619234085, -0.11233092099428177, 0.12257052958011627, -0.10201627016067505, 0.12511253356933594, -0.06524788588285446, 0.21065334975719452, 0.04297342896461487, -0.14576071500778198, 0.04701591655611992, -0.039328061044216156, -0.10480491816997528, -0.10834206640720367, -0.005855637136846781, -0.07763518393039703, -0.10801421105861664, 0.035854894667863846, -0.11808851361274719, 0.06373994797468185, 0.10413969308137894, 0.032328907400369644, 0.02847679890692234, 0.10821525007486343, -0.012818551622331142, 0.003967189230024815, 0.0615202933549881, 0.024190310388803482, -0.0013623898848891258, -0.06803473830223083, -0.07054542005062103, 0.03152662143111229, 0.03093055821955204, 0.08836309611797333, -0.041048310697078705, -0.015828432515263557, 0.04514778405427933, -0.011088612489402294, -0.07081957906484604, 0.03716809302568436, 0.002156134694814682, 0.06756417453289032, 0.04965738579630852, 0.05058959871530533, 0.014773912727832794, -0.04840933158993721, 0.29020994901657104, -0.06007274612784386, -0.11533622443675995, -0.12196926772594452, 0.23404274880886078, 0.05164040997624397, -0.02209814451634884, 0.06889677047729492, -0.09965430945158005, -0.05030576139688492, 0.16866910457611084, 0.16552342474460602, -0.08448578417301178, -0.02838929556310177, -0.02360529638826847, -0.012780903838574886, -0.03554441034793854, 0.14167630672454834, 0.11849255859851837, 0.0480002798140049, -0.05400398001074791, -0.028206059709191322, -0.023864882066845894, -0.03644191846251488, -0.06777505576610565, 0.0550810769200325, 0.021885735914111137, -0.004127000458538532, -0.0379343144595623, 0.06804746389389038, -0.012756816111505032, -0.20429426431655884, 0.023203959688544273, -0.16880296170711517, -0.18786080181598663, -0.008892958052456379, 0.09911592304706573, -0.029109714552760124, 0.04215659573674202, 0.009598146192729473, -0.008032025769352913, 0.09604009240865707, -0.03220779076218605, -0.020605554804205894, -0.10168667137622833, 0.06456213444471359, -0.12084256857633591, 0.23013290762901306, -0.006659603677690029, 0.07820934057235718, 0.09150468558073044, 0.030456386506557465, -0.1250225454568863, 0.06139509007334709, 0.04602694883942604, -0.10065801441669464, 0.03499612957239151, 0.15533967316150665, -0.044287052005529404, 0.042327940464019775, 0.03654909133911133, -0.12159112095832825, 0.002059836406260729, -0.08350995182991028, -0.029922865331172943, -0.05808687210083008, -0.019563153386116028, -0.036072783172130585, 0.149429589509964, 0.21446949243545532, -0.008592761121690273, 0.03154527768492699, -0.08619661629199982, -0.003914566244930029, 0.02025928720831871, 0.08992677181959152, -0.032664310187101364, -0.23212172091007233, 0.0337471179664135, 0.0205695740878582, 0.03170733153820038, -0.17975850403308868, -0.06468962877988815, 0.0056995814666152, -0.07137750089168549, -0.05807846039533615, 0.1114995926618576, 0.03764156997203827, 0.040624167770147324, -0.022844212129712105, -0.08261658996343613, -0.015339363366365433, 0.15085145831108093, -0.1842207908630371, -0.04368939995765686 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert_gpt2_summarization_cnndm_new This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "bert_gpt2_summarization_cnndm_new", "results": []}]}
text2text-generation
Ayham/bert_gpt2_summarization_cnndm_new
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# bert_gpt2_summarization_cnndm_new This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# bert_gpt2_summarization_cnndm_new\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# bert_gpt2_summarization_cnndm_new\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 62, 41, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# bert_gpt2_summarization_cnndm_new\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.08463947474956512, 0.1397019326686859, -0.0037230567540973425, 0.06574166566133499, 0.13799114525318146, 0.037261638790369034, 0.09401717782020569, 0.11788362264633179, -0.12208233773708344, 0.05834932625293732, 0.07389120012521744, 0.0664982795715332, 0.0483587421476841, 0.14604103565216064, -0.032225605100393295, -0.21457716822624207, 0.004854693077504635, -0.0013902181526646018, -0.07830501347780228, 0.11220652610063553, 0.08651818335056305, -0.10830793529748917, 0.06012361869215965, -0.001662667142227292, -0.1416771113872528, 0.01890520565211773, -0.04608428478240967, -0.0512525849044323, 0.09210645407438278, -0.010255131870508194, 0.09752532094717026, 0.03327793627977371, 0.14229515194892883, -0.20226646959781647, 0.0024338606745004654, 0.09578287601470947, 0.043176133185625076, 0.09239201992750168, 0.06384088099002838, -0.005158356856554747, 0.12380960583686829, -0.16194002330303192, 0.1002923920750618, 0.013413732871413231, -0.07597196847200394, -0.1205054223537445, -0.10079822689294815, 0.05784264951944351, 0.09159256517887115, 0.11824030429124832, 0.0066640120930969715, 0.12008292973041534, -0.11104154586791992, 0.07696489244699478, 0.16952039301395416, -0.2119087278842926, -0.04857321456074715, 0.061213962733745575, 0.02867666259407997, 0.06136168912053108, -0.09223591536283493, -0.009409269317984581, 0.02835116721689701, 0.023923885077238083, 0.09976688027381897, 0.007882663048803806, -0.1008763387799263, 0.005270147230476141, -0.12712812423706055, -0.02229159325361252, 0.08833150565624237, 0.019666556268930435, -0.01317521370947361, -0.09315042942762375, -0.04925587400794029, -0.14747190475463867, -0.026064850389957428, -0.022944536060094833, 0.04179415851831436, -0.06090404838323593, -0.06445756554603577, -0.05795867368578911, -0.06776636838912964, -0.05435852333903313, -0.023740505799651146, 0.1224903017282486, 0.04763033241033554, 0.00838720053434372, -0.046938907355070114, 0.12006592750549316, 0.057188376784324646, -0.11608700454235077, 0.0037256486248224974, 0.015857793390750885, -0.08938556909561157, -0.05552060902118683, -0.03459969907999039, -0.02855660393834114, -0.009688031859695911, 0.12882305681705475, -0.08124163001775742, 0.07805200666189194, 0.017192238941788673, -0.014368865638971329, -0.018122879788279533, 0.1407134234905243, -0.0521322637796402, -0.033852264285087585, -0.010427086614072323, 0.10762833803892136, 0.006206488236784935, -0.014308021403849125, -0.07228413969278336, -0.004140281118452549, 0.0690820962190628, 0.061670854687690735, -0.040487464517354965, 0.04621011018753052, -0.02497076243162155, -0.020412273705005646, 0.04214610159397125, -0.13667212426662445, 0.04275943338871002, 0.01426743995398283, -0.09914444386959076, -0.014412852004170418, 0.019500253722071648, 0.0021248681005090475, -0.049419526010751724, 0.1315208077430725, -0.07771699875593185, 0.00519027654081583, -0.08306682854890823, -0.07091567665338516, 0.01570930890738964, -0.10736659914255142, -0.03490092232823372, -0.03348951041698456, -0.18620282411575317, -0.05283847078680992, 0.02730599418282509, -0.05778954550623894, -0.04012368246912956, -0.059779584407806396, -0.07721900194883347, 0.02047692984342575, -0.008382447063922882, 0.1871756613254547, -0.06615234911441803, 0.06627131253480911, -0.026962239295244217, 0.023363541811704636, 0.05018841475248337, 0.04423291236162186, -0.06610596179962158, -0.0019317746628075838, -0.1049833670258522, 0.08939211815595627, -0.07952489703893661, 0.0035701566375792027, -0.10750197619199753, -0.0960516631603241, 0.014525213278830051, -0.011048912070691586, 0.05938521400094032, 0.1344480961561203, -0.19437000155448914, -0.034532900899648666, 0.11785310506820679, -0.047355037182569504, -0.015021352097392082, 0.05523981153964996, -0.054812490940093994, -0.010837869718670845, 0.06381478160619736, 0.13117532432079315, 0.06474532932043076, -0.130085289478302, -0.009991384111344814, 0.0016949439886957407, 0.03714430332183838, 0.005879279226064682, 0.022451452910900116, 0.001895626075565815, 0.054880619049072266, 0.004480496048927307, -0.04499338939785957, 0.002361156279221177, -0.09660527855157852, -0.07050664722919464, -0.03822485730051994, -0.08154080808162689, 0.022540001198649406, 0.0220720823854208, 0.03674738109111786, -0.058364734053611755, -0.11769036948680878, 0.09427841007709503, 0.12616465985774994, -0.055571604520082474, 0.017737695947289467, -0.0765766054391861, -0.023394037038087845, 0.005823349114507437, -0.023584501817822456, -0.1941407322883606, -0.11563108116388321, 0.02075890451669693, -0.06465738266706467, 0.038910072296857834, -0.011042634956538677, 0.06399063766002655, 0.044053249061107635, -0.043261587619781494, -0.01215632725507021, -0.06905616819858551, -0.00004012473436887376, -0.09456731379032135, -0.2077045440673828, -0.03181496635079384, -0.015461422502994537, 0.17005451023578644, -0.21854840219020844, 0.01385733112692833, -0.012989327311515808, 0.15075282752513885, 0.021462876349687576, -0.06166462600231171, 0.013100889511406422, 0.05075937509536743, 0.0073812128975987434, -0.089682437479496, 0.031534720212221146, -0.005823719315230846, -0.08263225108385086, -0.019207807257771492, -0.1347629427909851, -0.02694716490805149, 0.08241172879934311, 0.07690180093050003, -0.09954936802387238, 0.013492203317582607, -0.06980748474597931, -0.04314405843615532, -0.07732675969600677, 0.03223774582147598, 0.21275396645069122, 0.03074927255511284, 0.13397683203220367, -0.05340596288442612, -0.0842716172337532, -0.005653706379234791, 0.02840276248753071, 0.0057545555755496025, 0.09941032528877258, 0.09171853214502335, -0.04384567588567734, 0.07549808919429779, 0.029077306389808655, -0.04680744186043739, 0.13372626900672913, -0.042890116572380066, -0.08816570788621902, 0.001341945375315845, -0.021998144686222076, -0.02429727092385292, 0.10061372816562653, -0.05397871881723404, 0.007275640498846769, 0.03324839100241661, 0.021915404126048088, 0.042478010058403015, -0.184432253241539, 0.00024334172485396266, 0.01736489310860634, -0.05857371911406517, -0.03546483814716339, -0.02390538901090622, 0.054035164415836334, 0.10060268640518188, 0.01267531793564558, -0.02475002035498619, 0.01398705504834652, -0.002551600569859147, -0.06710200011730194, 0.1777106523513794, -0.11904040724039078, -0.16642433404922485, -0.08473005890846252, 0.03229621797800064, -0.03567327931523323, -0.04837607592344284, -0.002064503962174058, -0.11624528467655182, -0.06913221627473831, -0.1084912046790123, -0.018931210041046143, -0.00550834322348237, -0.0055945501662790775, 0.05720071867108345, 0.010167387314140797, 0.054009705781936646, -0.13599172234535217, 0.01603362336754799, -0.04671887308359146, -0.0833049863576889, 0.01140704844146967, 0.0846448764204979, 0.06418360024690628, 0.1496385782957077, -0.019401175901293755, 0.026183050125837326, -0.017759623005986214, 0.19132502377033234, -0.09707201272249222, 0.008085308596491814, 0.09951235353946686, 0.0119987977668643, 0.03981722891330719, 0.11240671575069427, 0.03814861550927162, -0.07936514168977737, 0.022942587733268738, 0.09495998919010162, -0.019930392503738403, -0.26391875743865967, -0.06812921911478043, -0.016604192554950714, -0.0854688361287117, 0.0847492441534996, 0.05404174327850342, -0.004471790511161089, 0.012665096670389175, -0.009274369105696678, -0.0022109481506049633, -0.01120303850620985, 0.05393917113542557, 0.08236098289489746, 0.05144631862640381, 0.0945046916604042, -0.03177529200911522, -0.026117661967873573, 0.06565895676612854, -0.0045778327621519566, 0.22317856550216675, -0.06224815174937248, 0.0813484713435173, 0.023524070158600807, 0.10221543908119202, -0.015186471864581108, 0.02515653893351555, 0.02408651076257229, -0.013403309509158134, 0.015868954360485077, -0.05212470516562462, -0.0016384921036660671, -0.00156576675362885, -0.009231863543391228, -0.003280953736975789, -0.06647537648677826, 0.03472612798213959, 0.014409272000193596, 0.2703491151332855, 0.0390915721654892, -0.29482319951057434, -0.05080012232065201, -0.019613783806562424, -0.025984948500990868, -0.05691303312778473, -0.007026399951428175, 0.09661538898944855, -0.11019323021173477, 0.07623066753149033, -0.04300127923488617, 0.08405294269323349, -0.06773363053798676, 0.0006079168524593115, 0.06960910558700562, 0.14358402788639069, -0.014231644570827484, 0.06258325278759003, -0.21612049639225006, 0.2121739238500595, 0.013560650870203972, 0.12322191148996353, -0.06915586441755295, 0.02201223187148571, 0.008457423187792301, 0.04686766117811203, 0.07157891243696213, -0.005456410348415375, -0.019716961309313774, -0.14918868243694305, -0.12326707690954208, 0.029485531151294708, 0.12215334177017212, -0.0010113166645169258, 0.08015413582324982, -0.019753722473978996, -0.013655148446559906, 0.040759146213531494, -0.15040050446987152, -0.1620301902294159, -0.12771263718605042, 0.03498243913054466, 0.03251970559358597, -0.041838765144348145, -0.054972127079963684, -0.11706238985061646, -0.03708278760313988, 0.18558771908283234, 0.014347556047141552, -0.059343770146369934, -0.14039723575115204, 0.06368265300989151, 0.15151196718215942, -0.03648693114519119, 0.019055521115660667, 0.032163601368665695, 0.11690769344568253, 0.04549073427915573, -0.08323396742343903, 0.03845752775669098, -0.06022057682275772, -0.18601535260677338, -0.049838680773973465, 0.13246403634548187, 0.08680745214223862, 0.04871742054820061, -0.01843336783349514, 0.028235286474227905, 0.005524190608412027, -0.0798250064253807, 0.008098862133920193, 0.08602108061313629, 0.09328653663396835, 0.06653455644845963, -0.06910540163516998, 0.0025991739239543676, -0.04141832888126373, -0.02659766748547554, 0.11075232923030853, 0.19284981489181519, -0.07851719856262207, 0.10475627332925797, 0.0652303695678711, -0.07476368546485901, -0.18333938717842102, 0.06565051525831223, 0.1277502179145813, 0.013648991473019123, 0.02501549944281578, -0.21054601669311523, 0.12116401642560959, 0.12526904046535492, -0.013228921219706535, 0.057324983179569244, -0.34221604466438293, -0.13233929872512817, 0.05212528631091118, 0.08399788290262222, 0.012243426404893398, -0.1042366549372673, -0.024020446464419365, -0.015882816165685654, -0.1382562220096588, 0.14687009155750275, -0.06587421149015427, 0.10812076181173325, -0.006694793235510588, 0.10538098961114883, 0.029598940163850784, -0.04815472289919853, 0.12069179862737656, 0.054013676941394806, 0.06858175247907639, -0.0497463122010231, 0.01895495131611824, 0.04220835864543915, -0.06561335176229477, 0.04075914993882179, -0.04136472940444946, 0.06906242668628693, -0.1190343052148819, -0.014410262927412987, -0.08652491122484207, 0.04394441470503807, -0.046554163098335266, -0.04240211844444275, -0.04469136521220207, 0.04800615459680557, 0.07050962001085281, -0.04421588033437729, 0.0784280002117157, 0.007615949492901564, 0.0873274952173233, 0.06244543567299843, 0.09702850878238678, -0.04792562127113342, -0.08123334497213364, -0.0025840471498668194, -0.013419832102954388, 0.03827780857682228, -0.11764974892139435, 0.026104846969246864, 0.1383369266986847, 0.048965658992528915, 0.1184963658452034, 0.036468952894210815, -0.04235413670539856, -0.020992591977119446, 0.03998963534832001, -0.12734933197498322, -0.10012982040643692, 0.02426786907017231, -0.0629180297255516, -0.10569245368242264, 0.020890705287456512, 0.10425086319446564, -0.03892531991004944, -0.01754414662718773, 0.0008252403349615633, 0.03000335767865181, 0.0039923409931361675, 0.20577526092529297, 0.02660459280014038, 0.06178769841790199, -0.11430063843727112, 0.12980642914772034, 0.07090382277965546, -0.09289708733558655, 0.054961323738098145, 0.11178263276815414, -0.09407392889261246, -0.010547167621552944, 0.09871666133403778, 0.14814816415309906, -0.03119531460106373, -0.03466680645942688, -0.08729484677314758, -0.09698494523763657, 0.08364195376634598, 0.12287812680006027, 0.023804981261491776, 0.005226412322372198, -0.047395214438438416, 0.019737910479307175, -0.15626290440559387, 0.06559370458126068, 0.0724731907248497, 0.05935705825686455, -0.09576108306646347, 0.1500711590051651, 0.02073080837726593, 0.03725479915738106, -0.018400657922029495, 0.019659696146845818, -0.0674453154206276, -0.009039957076311111, -0.09915674477815628, -0.0130859250202775, -0.008218938484787941, -0.004916087724268436, -0.01905549317598343, -0.04296484962105751, -0.04427057504653931, 0.042774587869644165, -0.07879003137350082, -0.0574883408844471, -0.0006984795327298343, 0.03366886079311371, -0.13617609441280365, 0.002866458846256137, -0.001152476412244141, -0.08988063037395477, 0.07271803170442581, 0.04634156450629234, 0.002611887874081731, 0.040634240955114365, -0.1422683745622635, -0.03911956027150154, 0.032982517033815384, 0.02191423997282982, 0.07594243437051773, -0.06716594099998474, -0.005733705125749111, -0.016076134517788887, 0.039138443768024445, 0.016009217128157616, 0.04637279361486435, -0.11006247252225876, 0.004013801459223032, -0.05702758952975273, -0.030613955110311508, -0.06796581298112869, 0.04904520511627197, 0.10703448951244354, 0.038548290729522705, 0.16774410009384155, -0.07568364590406418, 0.029540708288550377, -0.1768844723701477, -0.030643098056316376, -0.008889859542250633, -0.039492350071668625, -0.06058400124311447, -0.013731099665164948, 0.11124075204133987, -0.05445609614253044, 0.1297071874141693, 0.00544621841982007, 0.05955609306693077, 0.04005772992968559, -0.02707684598863125, -0.06499073654413223, 0.01985238306224346, 0.15159107744693756, 0.06533849239349365, -0.021644005551934242, 0.11270581185817719, -0.008844847790896893, 0.053038179874420166, 0.055052656680345535, 0.2281208634376526, 0.1290401667356491, 0.022717943415045738, 0.06959869712591171, 0.05146872624754906, -0.14472989737987518, -0.11751312017440796, 0.1178029477596283, -0.09579695016145706, 0.12881797552108765, -0.06404542922973633, 0.1981685906648636, 0.052641015499830246, -0.15014483034610748, 0.05337076261639595, -0.03372626751661301, -0.10803710669279099, -0.10874005407094955, -0.017886027693748474, -0.07805929332971573, -0.10314035415649414, 0.03225703537464142, -0.11551116406917572, 0.06357378512620926, 0.09647402912378311, 0.03412949666380882, 0.019903380423784256, 0.11887530237436295, -0.005692540667951107, -0.0017444441327825189, 0.06416048109531403, 0.023744333535432816, -0.0018626172095537186, -0.06898020952939987, -0.0702757015824318, 0.04008590802550316, 0.032304760068655014, 0.08631090819835663, -0.04007609561085701, -0.007474222686141729, 0.04830161854624748, -0.011036739684641361, -0.06937810778617859, 0.03932100534439087, 0.0036585561465471983, 0.06949752569198608, 0.05037375167012215, 0.0472574345767498, 0.009182692505419254, -0.049014441668987274, 0.2817234992980957, -0.06036731228232384, -0.1034669280052185, -0.11981377005577087, 0.22399134933948517, 0.05307126045227051, -0.02173003926873207, 0.07070163637399673, -0.10459437221288681, -0.04549245163798332, 0.15755370259284973, 0.17134703695774078, -0.06726491451263428, -0.025937266647815704, -0.026134846732020378, -0.010384496301412582, -0.032091837376356125, 0.12601225078105927, 0.10970599949359894, 0.05029814690351486, -0.04711693152785301, -0.023819709196686745, -0.014526769518852234, -0.031664539128541946, -0.07229869067668915, 0.06644581258296967, 0.015375633724033833, -0.006478387396782637, -0.03325515240430832, 0.0642036646604538, -0.0175995621830225, -0.20402924716472626, 0.030570292845368385, -0.17368188500404358, -0.1924954503774643, -0.012421614490449429, 0.09776695817708969, -0.027153218165040016, 0.041885603219270706, 0.015633463859558105, -0.014940910041332245, 0.10339691489934921, -0.030217640101909637, -0.03552204743027687, -0.10493657737970352, 0.06918732076883316, -0.11274181306362152, 0.23634617030620575, -0.0017128990730270743, 0.07177603989839554, 0.09094177186489105, 0.020741792395710945, -0.12725913524627686, 0.05505238473415375, 0.04664169251918793, -0.09068071097135544, 0.034172069281339645, 0.16438981890678406, -0.04737507551908493, 0.03816802427172661, 0.03457938879728317, -0.11278099566698074, -0.006244466174393892, -0.07919429987668991, -0.035440150648355484, -0.06315532326698303, -0.018697859719395638, -0.0386124886572361, 0.14913181960582733, 0.21492908895015717, -0.01016663946211338, 0.030581587925553322, -0.08016055822372437, 0.0024981889873743057, 0.022884663194417953, 0.09123212844133377, -0.02527233399450779, -0.22629472613334656, 0.03500283882021904, 0.008277487009763718, 0.038243528455495834, -0.18788860738277435, -0.06172402575612068, 0.012700226157903671, -0.06644535809755325, -0.05493411794304848, 0.10710149258375168, 0.05084388703107834, 0.03783103823661804, -0.02532951906323433, -0.07750063389539719, -0.01464747078716755, 0.1447814553976059, -0.18398071825504303, -0.0362023301422596 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert_gpt2_summarization_xsum This model is a fine-tuned version of [](https://huggingface.co/) on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["xsum"], "model-index": [{"name": "bert_gpt2_summarization_xsum", "results": []}]}
text2text-generation
Ayham/bert_gpt2_summarization_xsum
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:xsum", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us
# bert_gpt2_summarization_xsum This model is a fine-tuned version of [](URL on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# bert_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us \n", "# bert_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 59, 35, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us \n# bert_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.086587093770504, 0.14803096652030945, -0.002439304254949093, 0.05727662518620491, 0.14970242977142334, 0.04620835557579994, 0.09498182684183121, 0.11710729449987411, -0.09511981904506683, 0.08022240549325943, 0.0879121944308281, 0.06647380441427231, 0.06489133089780807, 0.14662031829357147, -0.03518614545464516, -0.25145912170410156, 0.02052290365099907, -0.02335074543952942, -0.0924774780869484, 0.09267972409725189, 0.08856693655252457, -0.11171279847621918, 0.0778687447309494, -0.002426356542855501, -0.16471680998802185, 0.016702737659215927, -0.047989413142204285, -0.05616242066025734, 0.09090844541788101, 0.010225331410765648, 0.09353043138980865, 0.016072828322649002, 0.12698723375797272, -0.22896210849285126, 0.0031129138078540564, 0.09078475087881088, 0.039289701730012894, 0.09221360832452774, 0.06297261267900467, 0.010586626827716827, 0.13671159744262695, -0.13806109130382538, 0.10629180818796158, 0.01656762696802616, -0.07124630361795425, -0.11999569088220596, -0.0855252742767334, 0.02033446915447712, 0.08340132236480713, 0.09092451632022858, 0.005923826713114977, 0.11534266918897629, -0.09419712424278259, 0.07163390517234802, 0.1662052720785141, -0.22922708094120026, -0.05761435255408287, 0.025774089619517326, 0.05882229655981064, 0.06697044521570206, -0.10459098219871521, -0.011427441611886024, 0.025909114629030228, 0.02617994137108326, 0.0818488821387291, -0.0001584346464369446, -0.09526284039020538, 0.013398810289800167, -0.125357523560524, -0.015895072370767593, 0.1230977475643158, 0.029753427952528, -0.027842381969094276, -0.10403180867433548, -0.05083340033888817, -0.11033394187688828, -0.01713622361421585, -0.031321894377470016, 0.04346070811152458, -0.03713296726346016, -0.07677487283945084, -0.043078456073999405, -0.06529263406991959, -0.06433742493391037, -0.019528072327375412, 0.1451338827610016, 0.03817478567361832, 0.012401100248098373, -0.0402684323489666, 0.09223975241184235, 0.04184785857796669, -0.11256680637598038, -0.0031826624181121588, -0.0038732949178665876, -0.10469502210617065, -0.04856949672102928, -0.06110748276114464, -0.03612678870558739, 0.012247951701283455, 0.13265222311019897, -0.059196192771196365, 0.10147170722484589, 0.025278490036725998, -0.013520128093659878, 0.00994619820266962, 0.14205177128314972, -0.04860362783074379, -0.04365392401814461, -0.023263322189450264, 0.07804654538631439, 0.01278948225080967, -0.022031325846910477, -0.0671645849943161, 0.004394868388772011, 0.07213862985372543, 0.05492362752556801, -0.05089510604739189, 0.026802198961377144, -0.04194853454828262, -0.025567300617694855, -0.0024201208725571632, -0.12488070130348206, 0.05233759805560112, 0.00005077346577309072, -0.08220219612121582, -0.006826815195381641, -0.0005854720366187394, 0.009891015477478504, -0.03236595541238785, 0.12340827286243439, -0.08881013840436935, -0.012411009520292282, -0.08876965939998627, -0.07712570577859879, 0.02659907564520836, -0.12964801490306854, -0.008838167414069176, -0.03793079033493996, -0.1663910299539566, -0.04400336742401123, 0.06362628936767578, -0.05994882807135582, -0.026819473132491112, -0.05317947268486023, -0.040274035185575485, 0.04169058799743652, -0.007703885436058044, 0.17661124467849731, -0.05818795785307884, 0.06508417427539825, -0.03485478088259697, 0.03883562609553337, 0.012639451771974564, 0.05218052864074707, -0.07047978043556213, 0.009043784812092781, -0.10740626603364944, 0.08349308371543884, -0.07059209048748016, 0.006595434155315161, -0.11705812066793442, -0.07896089553833008, -0.017077526077628136, -0.0015699099749326706, 0.07742748409509659, 0.1267833709716797, -0.19596020877361298, -0.03617032244801521, 0.11282740533351898, -0.0699540227651596, -0.057720717042684555, 0.06197666749358177, -0.050892528146505356, 0.010690857656300068, 0.04906652495265007, 0.16074378788471222, 0.04502710700035095, -0.11914761364459991, -0.030227813869714737, 0.01479552872478962, 0.05310356244444847, 0.01653832383453846, 0.050256527960300446, -0.013150586746633053, 0.03301861882209778, 0.01175924576818943, -0.02209498919546604, 0.007513976655900478, -0.07952336221933365, -0.074508436024189, -0.0351635180413723, -0.07354310154914856, -0.005566761828958988, 0.021185025572776794, 0.028071006760001183, -0.06058516353368759, -0.1026637852191925, 0.1117190420627594, 0.12284307926893234, -0.07530219107866287, 0.021345019340515137, -0.06402342021465302, -0.013535826466977596, -0.013104477897286415, -0.01788511499762535, -0.2031049281358719, -0.10754677653312683, 0.03436195105314255, -0.0629625990986824, 0.02693737857043743, -0.01580614596605301, 0.06040240824222565, 0.056513506919145584, -0.03169650211930275, -0.010445789434015751, -0.06741125881671906, -0.001224528648890555, -0.09856262803077698, -0.2175862044095993, -0.03848537802696228, -0.027019202709197998, 0.17956958711147308, -0.20508645474910736, -0.007081887219101191, -0.0009177476167678833, 0.13935448229312897, 0.030738383531570435, -0.07202717661857605, 0.001118972199037671, 0.04687250778079033, -0.009360921569168568, -0.09386395663022995, 0.034322187304496765, 0.011650417000055313, -0.09603915363550186, -0.036839280277490616, -0.14846745133399963, 0.00013207034498918802, 0.09544121474027634, 0.06324999779462814, -0.08440887928009033, -0.037802502512931824, -0.06440083682537079, -0.03878552466630936, -0.076694555580616, 0.018945591524243355, 0.1871514618396759, 0.016644535586237907, 0.11392951011657715, -0.06435221433639526, -0.07066063582897186, 0.01493808627128601, 0.03171633183956146, -0.02051933854818344, 0.09416419267654419, 0.12873421609401703, -0.08498028665781021, 0.08455043286085129, 0.07664316147565842, -0.031273987144231796, 0.14900891482830048, -0.0420624203979969, -0.08883573859930038, -0.006063756532967091, -0.004875252489000559, -0.022896505892276764, 0.11880844086408615, -0.09605821222066879, 0.01613294892013073, 0.027097661048173904, 0.03775639459490776, 0.05053296685218811, -0.17738279700279236, 0.008990178816020489, 0.02383304387331009, -0.03192060813307762, -0.05263494327664375, -0.020900102332234383, 0.025755152106285095, 0.08856803178787231, 0.02703297883272171, 0.007974631153047085, 0.014995359815657139, -0.005165532231330872, -0.07970178872346878, 0.19234396517276764, -0.1463257372379303, -0.1688394546508789, -0.09180039912462234, 0.03234610706567764, -0.04392470419406891, -0.030215730890631676, 0.008603495545685291, -0.1115613579750061, -0.06567800045013428, -0.09385671466588974, -0.019664492458105087, -0.052203159779310226, 0.013755766674876213, 0.044940318912267685, 0.010321445763111115, 0.045431435108184814, -0.13209910690784454, 0.013626650907099247, -0.05340103432536125, -0.06035284698009491, 0.0024323295801877975, 0.07340168952941895, 0.08875993639230728, 0.11670377850532532, -0.012959923595190048, 0.024812357500195503, -0.02842189557850361, 0.21274928748607635, -0.08269350975751877, -0.00959541741758585, 0.09908244013786316, -0.010791270062327385, 0.04566660150885582, 0.1123090386390686, 0.0325290821492672, -0.09486565738916397, 0.03287828713655472, 0.07057707756757736, -0.022487781941890717, -0.24646764993667603, -0.04727748781442642, -0.040235769003629684, -0.09014537930488586, 0.10520880669355392, 0.048496562987565994, -0.05674462392926216, 0.04154255986213684, -0.0005080584669485688, 0.031576208770275116, -0.026138294488191605, 0.061484530568122864, 0.07766447961330414, 0.04902108013629913, 0.09882982075214386, -0.024583155289292336, -0.01987374946475029, 0.075689397752285, 0.001094707171432674, 0.2644402086734772, -0.024470530450344086, 0.086273193359375, 0.028115056455135345, 0.09884726256132126, -0.01634308323264122, 0.060039274394512177, 0.015145063400268555, -0.007776259444653988, -0.009865890257060528, -0.05878129601478577, -0.037568941712379456, 0.02866704948246479, -0.005259022582322359, -0.00041826313827186823, -0.09217625856399536, 0.04844343662261963, 0.024715371429920197, 0.23634885251522064, 0.023087654262781143, -0.28616034984588623, -0.052097391337156296, -0.0007660966366529465, -0.029626356437802315, -0.0722634568810463, 0.0009631029097363353, 0.12399808317422867, -0.10855602473020554, 0.07219977676868439, -0.06333447247743607, 0.08423508703708649, -0.048125822097063065, -0.006548898294568062, 0.05914494767785072, 0.16110816597938538, -0.01262523140758276, 0.06549353897571564, -0.2066441774368286, 0.21446074545383453, 0.02481934055685997, 0.1205206885933876, -0.08114638179540634, 0.04558853060007095, 0.009703484363853931, 0.008577466942369938, 0.07953490316867828, 0.0007878708420321345, -0.11582661420106888, -0.14117464423179626, -0.08620886504650116, 0.06805748492479324, 0.15320754051208496, -0.015705378726124763, 0.07986229658126831, -0.03912361338734627, 0.013082440011203289, 0.04973310977220535, -0.08901788294315338, -0.16052721440792084, -0.16020001471042633, 0.025702739134430885, 0.01463642343878746, -0.031039483845233917, -0.05103790760040283, -0.09551553428173065, -0.02532264031469822, 0.16131635010242462, -0.00767900887876749, -0.0426512248814106, -0.15876521170139313, 0.07553119212388992, 0.1601378619670868, -0.05113900080323219, 0.027064520865678787, 0.01313087809830904, 0.1230865940451622, 0.035562071949243546, -0.09132643043994904, 0.06137733906507492, -0.07733874022960663, -0.15366442501544952, -0.06277958303689957, 0.10182949155569077, 0.06190093979239464, 0.04049627110362053, -0.015560051426291466, 0.04191455990076065, -0.009741955436766148, -0.09414112567901611, 0.031494949012994766, 0.09945053607225418, 0.07580581307411194, 0.06504514813423157, -0.09950210154056549, 0.04341130703687668, -0.007928828708827496, -0.030122997239232063, 0.13483378291130066, 0.1995646059513092, -0.0707472637295723, 0.10389278829097748, 0.09041603654623032, -0.08968312293291092, -0.19072723388671875, 0.08029474318027496, 0.1150459349155426, 0.02237631380558014, 0.0464707650244236, -0.24789321422576904, 0.13310661911964417, 0.1265239268541336, -0.004536381922662258, 0.05475641041994095, -0.293987512588501, -0.12543365359306335, 0.05501510202884674, 0.10602959245443344, 0.024326037615537643, -0.11963584274053574, -0.02383205108344555, -0.043410707265138626, -0.10960600525140762, 0.1564701646566391, -0.06984984874725342, 0.11325768381357193, 0.006821607705205679, 0.07796978950500488, 0.02467774972319603, -0.044299397617578506, 0.12858456373214722, 0.02140733413398266, 0.06406933814287186, -0.032010309398174286, 0.01786811649799347, 0.0010440570767968893, -0.05724750831723213, 0.03003130480647087, -0.08886181563138962, 0.06276743859052658, -0.11113423854112625, -0.016704227775335312, -0.0488329716026783, 0.0526595301926136, -0.0408441424369812, -0.058377038687467575, -0.03959104046225548, 0.04285462200641632, 0.07900165021419525, -0.0348360501229763, 0.07197420299053192, 0.0115796634927392, 0.0754479393362999, 0.04341490566730499, 0.10090004652738571, -0.0649474486708641, -0.054566022008657455, 0.00618396420031786, -0.012685486115515232, 0.053630102425813675, -0.12077422440052032, 0.025553904473781586, 0.13214242458343506, 0.04196607321500778, 0.13250142335891724, 0.05609156936407089, -0.044458549469709396, -0.002476498018950224, 0.05345219001173973, -0.11938994377851486, -0.12338273972272873, 0.0059024859219789505, -0.03147972375154495, -0.1273096799850464, 0.014645339921116829, 0.1032741516828537, -0.029780738055706024, -0.006428875494748354, -0.017258184030652046, 0.03875354677438736, 0.005983958952128887, 0.19159142673015594, 0.009549702517688274, 0.05280788615345955, -0.10755828022956848, 0.1450437754392624, 0.05640279874205589, -0.11476565897464752, 0.0761483758687973, 0.08146706968545914, -0.08381550014019012, 0.0015300854574888945, 0.06625379621982574, 0.14100107550621033, -0.02478770911693573, -0.04906927049160004, -0.08474461734294891, -0.09607767313718796, 0.06145690754055977, 0.1005370244383812, 0.025923119857907295, 0.005381923634558916, -0.04970334842801094, 0.03565150499343872, -0.1510869711637497, 0.0657208040356636, 0.04851958528161049, 0.06453659385442734, -0.10674543678760529, 0.10725999623537064, 0.018812431022524834, 0.010003570467233658, -0.013072188943624496, 0.014731097035109997, -0.1021023765206337, -0.02918764017522335, -0.07837153226137161, -0.014065489172935486, -0.03364940360188484, -0.0033222129568457603, -0.0014552893117070198, -0.03326137736439705, -0.05590049922466278, 0.03282973915338516, -0.07885589450597763, -0.059211570769548416, -0.0072306860238313675, 0.043069981038570404, -0.1354685127735138, 0.009997008368372917, 0.01660575345158577, -0.10202021896839142, 0.0759044736623764, 0.05634661391377449, 0.020986244082450867, 0.037685226649045944, -0.14511534571647644, -0.041577234864234924, 0.030936401337385178, 0.021246885880827904, 0.0658356174826622, -0.08940795063972473, -0.00894484855234623, -0.006287492346018553, 0.041872404515743256, 0.00715664541348815, 0.0636478140950203, -0.1129119023680687, -0.0054219067096710205, -0.07233721017837524, -0.05306072160601616, -0.062035199254751205, 0.04323333874344826, 0.10887487232685089, 0.04094203561544418, 0.1630237102508545, -0.061172425746917725, 0.02693810500204563, -0.18022672832012177, -0.022452063858509064, -0.004666696302592754, -0.03632878512144089, -0.08097270876169205, -0.024984125047922134, 0.0926244854927063, -0.05044049397110939, 0.1393839567899704, 0.008350677788257599, 0.05761506035923958, 0.03420117869973183, -0.021636368706822395, -0.04283907637000084, 0.009884970262646675, 0.20156782865524292, 0.08297409117221832, -0.019312405958771706, 0.08769398182630539, 0.006522571202367544, 0.067653588950634, 0.06819549947977066, 0.22393043339252472, 0.1458996832370758, 0.0003200028440915048, 0.08813386410474777, 0.05339302495121956, -0.13237258791923523, -0.11442545801401138, 0.15819571912288666, -0.059867966920137405, 0.1291087418794632, -0.04554911330342293, 0.19307686388492584, 0.05440036579966545, -0.17186400294303894, 0.03096025064587593, -0.04195791855454445, -0.10218454152345657, -0.13269956409931183, -0.005197884049266577, -0.0865531638264656, -0.11740200221538544, 0.024851659312844276, -0.13244281709194183, 0.05476374924182892, 0.08132722228765488, 0.018806619569659233, 0.03658527880907059, 0.11151216179132462, -0.03610733896493912, 0.0013915011659264565, 0.05032338574528694, 0.03193189203739166, -0.0015339170349761844, -0.0491972453892231, -0.0931544080376625, 0.03385043144226074, 0.01207559835165739, 0.09250126779079437, -0.04940143972635269, -0.0111940186470747, 0.05418921634554863, -0.008450784720480442, -0.06157034635543823, 0.017646552994847298, 0.0038464698009192944, 0.039690542966127396, 0.04126505181193352, 0.04571124538779259, -0.005138057749718428, -0.04553154855966568, 0.26492801308631897, -0.06674668192863464, -0.054287686944007874, -0.13547715544700623, 0.19507238268852234, 0.0422392413020134, -0.006575934588909149, 0.07241123169660568, -0.10694355517625809, -0.049592792987823486, 0.16529664397239685, 0.15348224341869354, -0.07736832648515701, -0.03250956907868385, -0.004780881106853485, -0.009718143381178379, -0.04985587298870087, 0.14544492959976196, 0.0967404916882515, 0.06838337332010269, -0.049399930983781815, -0.02081379108130932, -0.019740669056773186, -0.0452587828040123, -0.057992033660411835, 0.06097623333334923, 0.025814233347773552, -0.015394657850265503, -0.025732040405273438, 0.06739359349012375, -0.009975501336157322, -0.1905539482831955, 0.02649928256869316, -0.17328864336013794, -0.17669346928596497, -0.020924808457493782, 0.09795330464839935, -0.027026692405343056, 0.050028298050165176, -0.001634739339351654, -0.01369502954185009, 0.10069427639245987, -0.005638578440994024, -0.05660254508256912, -0.10683540254831314, 0.07792934030294418, -0.07142136991024017, 0.21960169076919556, -0.004257912281900644, 0.059583064168691635, 0.09966909885406494, 0.0534830242395401, -0.12723180651664734, 0.025927118957042694, 0.059200718998909, -0.10587280988693237, 0.03686083108186722, 0.143033966422081, -0.06693337112665176, 0.07422613352537155, 0.04538920149207115, -0.09757914394140244, 0.0020086318254470825, -0.07516726106405258, -0.039460115134716034, -0.04978456720709801, -0.016401011496782303, -0.07092674821615219, 0.16515183448791504, 0.21078704297542572, -0.013323978520929813, 0.010404940694570541, -0.07505428045988083, 0.04131850227713585, 0.031324226409196854, 0.09599780291318893, -0.0316317118704319, -0.2349216490983963, 0.017095476388931274, 0.03580310568213463, 0.016107484698295593, -0.20204752683639526, -0.07304736971855164, 0.02990403212606907, -0.0490150973200798, -0.06463830173015594, 0.09957960993051529, 0.06188155338168144, 0.023206761106848717, -0.04515459015965462, -0.10389596968889236, -0.0216011181473732, 0.13917148113250732, -0.1439414769411087, -0.04150782898068428 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert_roberta_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "bert_roberta_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/bert_roberta_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# bert_roberta_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
[ "# bert_roberta_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# bert_roberta_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ 62, 40, 6, 12, 8, 3, 118, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# bert_roberta_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ -0.08107836544513702, 0.13498660922050476, -0.003843706101179123, 0.06884738057851791, 0.14027850329875946, 0.03641236200928688, 0.10311071574687958, 0.11609242111444473, -0.11337988823652267, 0.055602576583623886, 0.07278820127248764, 0.06978823989629745, 0.04607595130801201, 0.157814159989357, -0.03085101954638958, -0.2204674929380417, 0.0005722203059121966, -0.008483171463012695, -0.08171654492616653, 0.1165754497051239, 0.08409684896469116, -0.10326474905014038, 0.059823550283908844, -0.002236961852759123, -0.13445588946342468, 0.021459633484482765, -0.04300154000520706, -0.05215442180633545, 0.09558900445699692, -0.0023775335866957903, 0.09544351696968079, 0.031886983662843704, 0.1472279578447342, -0.20103928446769714, 0.004880353808403015, 0.10045254230499268, 0.04024586081504822, 0.0957309752702713, 0.05977090075612068, -0.013683188706636429, 0.1323520988225937, -0.15068626403808594, 0.10300332307815552, 0.01884547621011734, -0.07754316180944443, -0.12155573070049286, -0.10158959031105042, 0.05212797224521637, 0.09091576188802719, 0.11799860000610352, 0.0064469208009541035, 0.1318984180688858, -0.1086912676692009, 0.08163139969110489, 0.17980989813804626, -0.20661433041095734, -0.04781162366271019, 0.060636840760707855, 0.023308245465159416, 0.05464734882116318, -0.09454891830682755, -0.017539629712700844, 0.023257199674844742, 0.022533433511853218, 0.1016852855682373, 0.0027838030364364386, -0.10102436691522598, -0.0005501307314261794, -0.12558650970458984, -0.01909445971250534, 0.07376188784837723, 0.022504881024360657, -0.010562572628259659, -0.09285929054021835, -0.0565505288541317, -0.1409272998571396, -0.02451745793223381, -0.00938098132610321, 0.03878531977534294, -0.056443773210048676, -0.06844452023506165, -0.041774995625019073, -0.06546419113874435, -0.04811893403530121, -0.025267239660024643, 0.12486357241868973, 0.04700640216469765, 0.008625910617411137, -0.042673613876104355, 0.11383925378322601, 0.07038891315460205, -0.12199476361274719, 0.013126788660883904, 0.014629971235990524, -0.09583450108766556, -0.05068407952785492, -0.02612425945699215, -0.024958571419119835, -0.009241250343620777, 0.1317891627550125, -0.07362368702888489, 0.0811104029417038, 0.02002791129052639, -0.015736663714051247, -0.017503084614872932, 0.14656245708465576, -0.04928245022892952, -0.03868158161640167, -0.010865430347621441, 0.10824324935674667, 0.006244024261832237, -0.013066543266177177, -0.06907966732978821, 0.011405942030251026, 0.06281084567308426, 0.06213284656405449, -0.042248260229825974, 0.047001197934150696, -0.01980314962565899, -0.01545295026153326, 0.04501562938094139, -0.13628624379634857, 0.04122872278094292, 0.01802440546452999, -0.10185118764638901, -0.013124573975801468, 0.016470452770590782, -0.004473160021007061, -0.04777141660451889, 0.12689709663391113, -0.07723452895879745, 0.007462698966264725, -0.08277343958616257, -0.08114732056856155, 0.012966428883373737, -0.09706476330757141, -0.03866782784461975, -0.035380445420742035, -0.17591381072998047, -0.05871058627963066, 0.028832977637648582, -0.06150590255856514, -0.03642977401614189, -0.05526768043637276, -0.07741845399141312, 0.023000018671154976, -0.006777926813811064, 0.1843370795249939, -0.06889589130878448, 0.06722947210073471, -0.02392628788948059, 0.029585877433419228, 0.06582491844892502, 0.04747871309518814, -0.07074618339538574, -0.009523390792310238, -0.10235251486301422, 0.08703181892633438, -0.0822427049279213, 0.007158446125686169, -0.10989689081907272, -0.09349644184112549, 0.02593693882226944, -0.008520171977579594, 0.056053806096315384, 0.12893876433372498, -0.19579653441905975, -0.035392191261053085, 0.1138404905796051, -0.052658356726169586, -0.005827719811350107, 0.05107535421848297, -0.05540851503610611, -0.007623913232237101, 0.06280023604631424, 0.12154767662286758, 0.06537721306085587, -0.1229361742734909, -0.004789162427186966, 0.0015872146468609571, 0.03482160344719887, 0.00986446999013424, 0.017855901271104813, 0.004719162825495005, 0.05906637758016586, 0.006251022685319185, -0.03571410849690437, 0.011291109025478363, -0.09667990356683731, -0.06504885852336884, -0.03308960050344467, -0.07492563128471375, 0.0163392573595047, 0.030552593991160393, 0.03363819792866707, -0.05782151222229004, -0.11822403967380524, 0.08385799080133438, 0.12412784993648529, -0.055824484676122665, 0.01563028059899807, -0.07607555389404297, -0.03797925263643265, 0.011555662378668785, -0.0186042170971632, -0.19924584031105042, -0.11612758040428162, 0.020264098420739174, -0.06518112868070602, 0.04168684035539627, -0.008459402248263359, 0.06414956599473953, 0.04711346700787544, -0.037917446345090866, -0.01632419042289257, -0.060278475284576416, 0.003152080811560154, -0.09153423458337784, -0.2124248445034027, -0.032558076083660126, -0.018282601609826088, 0.1561003178358078, -0.21931906044483185, 0.012582255527377129, -0.01881362684071064, 0.14640873670578003, 0.026290412992239, -0.05728372931480408, 0.016395170241594315, 0.05375680699944496, 0.006485589314252138, -0.09339331090450287, 0.03250476345419884, -0.010673388838768005, -0.07360813021659851, -0.012212743982672691, -0.14051401615142822, -0.026080157607793808, 0.0796126201748848, 0.09095526486635208, -0.10678602755069733, 0.025100285187363625, -0.07332081347703934, -0.04099142551422119, -0.09010941535234451, 0.03752158209681511, 0.20622923970222473, 0.036258772015571594, 0.1347283571958542, -0.046174418181180954, -0.08453484624624252, -0.007972624152898788, 0.02135240100324154, 0.01248570904135704, 0.10921502858400345, 0.08002030849456787, -0.04549550265073776, 0.07451357692480087, 0.01788449101150036, -0.04105379059910774, 0.13080884516239166, -0.04437609761953354, -0.086020827293396, 0.0033259624615311623, -0.029818838462233543, -0.023361951112747192, 0.09827503561973572, -0.03204336017370224, 0.005203686188906431, 0.03312243893742561, 0.02445930801331997, 0.03144148364663124, -0.18450751900672913, 0.0011510658077895641, 0.014994543977081776, -0.05631338059902191, -0.03924601525068283, -0.02042962983250618, 0.05570618435740471, 0.10396440327167511, 0.005511854775249958, -0.035269152373075485, 0.008427047170698643, -0.006207200698554516, -0.06799686700105667, 0.17724785208702087, -0.11518634110689163, -0.16893449425697327, -0.08150651305913925, 0.015118780545890331, -0.02826118841767311, -0.050443656742572784, 0.0032767001539468765, -0.11885134875774384, -0.07419691234827042, -0.11383984237909317, -0.012576421722769737, 0.011755507439374924, -0.0020813627634197474, 0.04531463235616684, 0.014302975498139858, 0.05213048681616783, -0.13858100771903992, 0.01516256108880043, -0.051931045949459076, -0.0787658542394638, 0.018882649019360542, 0.0814734399318695, 0.05989126116037369, 0.14927276968955994, -0.01649605482816696, 0.027649205178022385, -0.017168765887618065, 0.19229917228221893, -0.09204304963350296, 0.012805513106286526, 0.09999848902225494, 0.015247217379510403, 0.04374711960554123, 0.1236623153090477, 0.041350677609443665, -0.0760091245174408, 0.029265139251947403, 0.09836433082818985, -0.024324623867869377, -0.25711116194725037, -0.06499703228473663, -0.016471412032842636, -0.08279350399971008, 0.08361529558897018, 0.055848997086286545, -0.01590975560247898, 0.004092933610081673, -0.007906106300652027, -0.015282156877219677, -0.004336596000939608, 0.052012231200933456, 0.07901857793331146, 0.05497773736715317, 0.09365958720445633, -0.034253571182489395, -0.027691390365362167, 0.06422823667526245, -0.006723711267113686, 0.22332841157913208, -0.06542927771806717, 0.07192967087030411, 0.021022483706474304, 0.08631417900323868, -0.012725966982543468, 0.02896457351744175, 0.02343592792749405, -0.01479268353432417, 0.012720366939902306, -0.04714673385024071, 0.0014905905118212104, -0.00010951447620755062, -0.01445078570395708, -0.0036950481589883566, -0.06367218494415283, 0.03567245230078697, 0.013923563063144684, 0.2786695063114166, 0.03911107778549194, -0.2877774238586426, -0.051197219640016556, -0.023367902263998985, -0.03468235582113266, -0.046214643865823746, -0.01014663651585579, 0.09207828342914581, -0.11416566371917725, 0.06545208394527435, -0.04386202245950699, 0.08410274982452393, -0.07525937259197235, -0.0016175875207409263, 0.06602153182029724, 0.1479947566986084, -0.017917895689606667, 0.06016678363084793, -0.21252310276031494, 0.2225715070962906, 0.007627885788679123, 0.12250969558954239, -0.06377160549163818, 0.0199875645339489, 0.016917860135436058, 0.038695722818374634, 0.06315286457538605, -0.006312972400337458, -0.023884931579232216, -0.16144174337387085, -0.11884666234254837, 0.024557610973715782, 0.11385598033666611, 0.0014140690909698606, 0.08260615170001984, -0.025234252214431763, -0.018677404150366783, 0.03899959847331047, -0.15944215655326843, -0.14981147646903992, -0.12215840071439743, 0.03107597306370735, 0.028821857646107674, -0.04136579483747482, -0.05617181584239006, -0.11866553127765656, -0.02353055588901043, 0.17476873099803925, 0.027605917304754257, -0.06092458963394165, -0.1455356329679489, 0.06375952064990997, 0.14837166666984558, -0.04278422147035599, 0.009670468978583813, 0.03240250423550606, 0.11851517111063004, 0.04896807298064232, -0.08700310438871384, 0.03840664401650429, -0.06254572421312332, -0.18770821392536163, -0.05597648024559021, 0.12841208279132843, 0.09052792936563492, 0.050394028425216675, -0.022529639303684235, 0.029392564669251442, 0.0016300692223012447, -0.07694092392921448, 0.006971640512347221, 0.08252204209566116, 0.0759914442896843, 0.06075360253453255, -0.06712254136800766, -0.005330916494131088, -0.04376021772623062, -0.02155659906566143, 0.10386889427900314, 0.19187869131565094, -0.0783262774348259, 0.11080913245677948, 0.06711333245038986, -0.06663298606872559, -0.17992813885211945, 0.05647554248571396, 0.12785738706588745, 0.00881000142544508, 0.03137799724936485, -0.205485999584198, 0.12301861494779587, 0.11875279992818832, -0.011682610027492046, 0.04955684021115303, -0.3465808033943176, -0.12558278441429138, 0.05890600010752678, 0.08400332927703857, 0.01879403553903103, -0.10508027672767639, -0.01933443173766136, -0.015114863403141499, -0.13317756354808807, 0.14392505586147308, -0.0581531897187233, 0.1115124523639679, -0.007817764766514301, 0.10356778651475906, 0.024189261719584465, -0.04292675107717514, 0.10350018739700317, 0.058141887187957764, 0.06972279399633408, -0.04616670310497284, 0.020415544509887695, 0.03830119967460632, -0.06392273306846619, 0.03717181459069252, -0.04515799880027771, 0.06819744408130646, -0.11777965724468231, -0.012543759308755398, -0.08517105877399445, 0.04381817206740379, -0.041869983077049255, -0.04307937249541283, -0.04050615057349205, 0.05288860201835632, 0.07494261115789413, -0.04746624454855919, 0.07144015282392502, 0.004945692140609026, 0.08681098371744156, 0.06761961430311203, 0.0859735831618309, -0.05096946284174919, -0.092528335750103, -0.006363812368363142, -0.009971430525183678, 0.03237579017877579, -0.11780504137277603, 0.0241203922778368, 0.13326190412044525, 0.049195434898138046, 0.1203073188662529, 0.034855980426073074, -0.03868116810917854, -0.017960067838430405, 0.037762679159641266, -0.11917967349290848, -0.08960983157157898, 0.03162027522921562, -0.06032847240567207, -0.10613039135932922, 0.014124799519777298, 0.10274393856525421, -0.03510443866252899, -0.020069772377610207, 0.004409019835293293, 0.031468894332647324, 0.014627041295170784, 0.21654121577739716, 0.028065156191587448, 0.06578309088945389, -0.11726891249418259, 0.12896770238876343, 0.06373598426580429, -0.08538620173931122, 0.04734526202082634, 0.13036558032035828, -0.10039278119802475, -0.011736493557691574, 0.10356897115707397, 0.1398620754480362, -0.034342873841524124, -0.02779916487634182, -0.0926666110754013, -0.09481994062662125, 0.07999736815690994, 0.13500022888183594, 0.030768664553761482, 0.0014475652715191245, -0.053067632019519806, 0.013646823354065418, -0.15053190290927887, 0.06761212646961212, 0.0816536545753479, 0.05453560873866081, -0.08123205602169037, 0.13633112609386444, 0.026367396116256714, 0.03449748083949089, -0.018526166677474976, 0.013501928187906742, -0.07453111559152603, -0.010938312858343124, -0.08670385181903839, -0.01449705008417368, -0.009742594324052334, -0.003427839372307062, -0.022213544696569443, -0.049846503883600235, -0.03938744217157364, 0.04453163594007492, -0.07786353677511215, -0.05890174210071564, -0.0020974718499928713, 0.0394115075469017, -0.13934677839279175, 0.001729165785945952, 0.0043597714975476265, -0.0921102911233902, 0.06880898028612137, 0.04396073520183563, -0.004285789094865322, 0.034205589443445206, -0.15396073460578918, -0.04497584328055382, 0.03171677142381668, 0.014557055197656155, 0.08113010227680206, -0.07779011130332947, -0.0015746894059702754, -0.02163849025964737, 0.03913373500108719, 0.014096572995185852, 0.053854092955589294, -0.11294689774513245, 0.007564733270555735, -0.06995955109596252, -0.03283659741282463, -0.06894779205322266, 0.04249308630824089, 0.10685157775878906, 0.043921079486608505, 0.17710013687610626, -0.08017106354236603, 0.037831153720617294, -0.1785109043121338, -0.032081592828035355, -0.0052351211197674274, -0.0430900938808918, -0.06879160553216934, -0.013873148709535599, 0.10949883610010147, -0.05326760932803154, 0.12084125727415085, 0.014042900875210762, 0.06973893195390701, 0.04213889688253403, -0.026322444900870323, -0.06305214762687683, 0.014452025294303894, 0.13918189704418182, 0.06137939542531967, -0.020318809896707535, 0.11476283520460129, -0.009758089669048786, 0.050620149821043015, 0.04612843319773674, 0.22578659653663635, 0.1285749226808548, 0.027547849342226982, 0.06811875849962234, 0.04808546230196953, -0.14300329983234406, -0.1177237257361412, 0.10790538042783737, -0.09396789968013763, 0.12739409506320953, -0.06629768759012222, 0.2103004902601242, 0.05276104435324669, -0.14340199530124664, 0.05475129932165146, -0.030494702979922295, -0.10121653228998184, -0.10564501583576202, -0.00833465252071619, -0.0781322717666626, -0.10775217413902283, 0.03636791184544563, -0.11066645383834839, 0.06661885231733322, 0.10212775319814682, 0.029861396178603172, 0.0296628475189209, 0.11134043335914612, -0.006428322289139032, -0.0018203752115368843, 0.0572904497385025, 0.024821819737553596, -0.00472657335922122, -0.06509522348642349, -0.0627424493432045, 0.027482779696583748, 0.02547461912035942, 0.08254890888929367, -0.0340469554066658, -0.0077611966989934444, 0.04672752320766449, -0.01992570236325264, -0.07643155753612518, 0.03989766165614128, 0.0037561405915766954, 0.06730307638645172, 0.056277673691511154, 0.056021831929683685, 0.007196280639618635, -0.0480811633169651, 0.29041415452957153, -0.06370855867862701, -0.10786508768796921, -0.12468152493238449, 0.24034175276756287, 0.05723869428038597, -0.023443983867764473, 0.0625985860824585, -0.10406484454870224, -0.0489320382475853, 0.15840759873390198, 0.16721884906291962, -0.05713550001382828, -0.02651510015130043, -0.03374389186501503, -0.012689530849456787, -0.028697948902845383, 0.12807314097881317, 0.11462491005659103, 0.052031729370355606, -0.050205882638692856, -0.02466502971947193, -0.01893633045256138, -0.0411149300634861, -0.07319090515375137, 0.07161898165941238, 0.021753927692770958, -0.004930845927447081, -0.03662196174263954, 0.06882565468549728, -0.011746928095817566, -0.21409396827220917, 0.03554947301745415, -0.1675889492034912, -0.185629203915596, -0.008736835792660713, 0.09460695087909698, -0.022337263450026512, 0.04726800322532654, 0.0169221218675375, -0.01240864023566246, 0.10283365845680237, -0.029831470921635628, -0.03342374414205551, -0.10812856256961823, 0.06506345421075821, -0.13239243626594543, 0.2291552573442459, -0.007560335099697113, 0.06465142220258713, 0.09346146136522293, 0.021842800080776215, -0.12682634592056274, 0.06597713381052017, 0.04461095854640007, -0.09956099838018417, 0.031989194452762604, 0.16407743096351624, -0.042603231966495514, 0.04542357102036476, 0.03725844621658325, -0.11739794909954071, -0.0014877445064485073, -0.09210249036550522, -0.032458461821079254, -0.05780152976512909, -0.017410479485988617, -0.029163740575313568, 0.14812345802783966, 0.2185637205839157, -0.00797269307076931, 0.029252631589770317, -0.08252638578414917, -0.0063522979617118835, 0.024208784103393555, 0.07745891809463501, -0.029338127002120018, -0.22232423722743988, 0.03354167193174362, 0.009722452610731125, 0.03758979216217995, -0.18420347571372986, -0.06988559663295746, 0.003587693441659212, -0.06968852877616882, -0.05222427099943161, 0.11014220863580704, 0.04538045823574066, 0.041354816406965256, -0.024778783321380615, -0.077470563352108, -0.010349351912736893, 0.14737291634082794, -0.19048988819122314, -0.03736972436308861 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bertgpt2_cnn This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "bertgpt2_cnn", "results": []}]}
text2text-generation
Ayham/bertgpt2_cnn
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
# bertgpt2_cnn This model is a fine-tuned version of [](URL on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# bertgpt2_cnn\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n", "# bertgpt2_cnn\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 53, 31, 6, 12, 8, 3, 118, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n# bertgpt2_cnn\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.0713464617729187, 0.13724267482757568, -0.003409226890653372, 0.06122308224439621, 0.14158909022808075, 0.03485475108027458, 0.10631176084280014, 0.12324313074350357, -0.0886872410774231, 0.0868392214179039, 0.08769725263118744, 0.06013815850019455, 0.05977007374167442, 0.11183979362249374, -0.031337399035692215, -0.25420665740966797, 0.030060991644859314, 0.004420667886734009, -0.0809573382139206, 0.09159436076879501, 0.09562042355537415, -0.09871209412813187, 0.06852208822965622, 0.011991473846137524, -0.15204203128814697, 0.019875120371580124, -0.05801241472363472, -0.06615182012319565, 0.07981996983289719, 0.02808472327888012, 0.09672998636960983, 0.022553039714694023, 0.09613872319459915, -0.2347533404827118, 0.005834112875163555, 0.06814893335103989, 0.02708500251173973, 0.08879697322845459, 0.07680948823690414, -0.0007602287805639207, 0.11511079967021942, -0.14240828156471252, 0.10367771238088608, 0.028027042746543884, -0.058684639632701874, -0.13831806182861328, -0.0789879783987999, 0.07238371670246124, 0.0754089504480362, 0.08440675586462021, 0.007227918598800898, 0.1267595887184143, -0.10462242364883423, 0.07393856346607208, 0.19735242426395416, -0.26068317890167236, -0.05606025829911232, 0.019902193918824196, 0.06934864073991776, 0.06298898160457611, -0.09935003519058228, -0.009389769285917282, 0.039806753396987915, 0.033220868557691574, 0.08716510236263275, 0.007184233982115984, -0.06001615151762962, -0.0014258784940466285, -0.13178220391273499, -0.02626875229179859, 0.14362329244613647, 0.053431734442710876, -0.035742007195949554, -0.12277215719223022, -0.041416481137275696, -0.10321912169456482, -0.02140413410961628, -0.03502289950847626, 0.03587980940937996, -0.03390311822295189, -0.07264313846826553, -0.06319601088762283, -0.06958208233118057, -0.05941227450966835, -0.006203522905707359, 0.16606661677360535, 0.059960346668958664, -0.004474806599318981, -0.03348829597234726, 0.0951925739645958, 0.06161453574895859, -0.11419444531202316, -0.015730703249573708, -0.0062186759896576405, -0.07083036005496979, -0.05195377394556999, -0.0626375824213028, -0.05509958043694496, 0.025266757234930992, 0.15202106535434723, -0.06035776063799858, 0.09992988407611847, 0.0071683358401060104, 0.008806279860436916, -0.0022332591470330954, 0.12318795174360275, -0.04687055945396423, -0.006178040523082018, -0.015222439542412758, 0.09193283319473267, 0.012072641402482986, -0.022369656711816788, -0.09055473655462265, -0.027010181918740273, 0.08926329761743546, 0.061645474284887314, -0.037626203149557114, 0.035046402364969254, -0.056078750640153885, -0.0310598686337471, 0.012970873154699802, -0.1203746423125267, 0.046485785394907, 0.013847527094185352, -0.07819342613220215, 0.025375844910740852, 0.029213514178991318, 0.004291544668376446, -0.046799223870038986, 0.08395202457904816, -0.08209613710641861, -0.010073278099298477, -0.07320738583803177, -0.067232646048069, 0.03452521562576294, -0.0627603754401207, -0.008028818294405937, -0.07908286154270172, -0.1430143266916275, -0.03391820564866066, 0.04208790883421898, -0.054065339267253876, -0.035352300852537155, -0.05285424739122391, -0.04856036603450775, 0.04253910109400749, -0.01307208277285099, 0.12540583312511444, -0.0503142848610878, 0.07112272083759308, -0.015513046644628048, 0.035797763615846634, 0.02355332486331463, 0.04624268412590027, -0.06319653242826462, 0.02635364420711994, -0.11545160412788391, 0.09178171306848526, -0.0964125245809555, 0.008571972139179707, -0.12112437933683395, -0.09701980650424957, -0.0070749083533883095, -0.003461043117567897, 0.08065830916166306, 0.1328195184469223, -0.1245206743478775, -0.04705234616994858, 0.14201119542121887, -0.06110665202140808, -0.06655548512935638, 0.09593778103590012, -0.04193442314863205, -0.012512694112956524, 0.05106412619352341, 0.136330246925354, 0.08433583378791809, -0.1155497208237648, -0.030156495049595833, 0.028136659413576126, 0.05959033593535423, 0.021760769188404083, 0.06842108070850372, -0.013699855655431747, -0.005665895529091358, 0.01880507916212082, -0.026199527084827423, 0.023215211927890778, -0.0861600786447525, -0.0789702758193016, -0.047483816742897034, -0.07476827502250671, 0.053690869361162186, 0.005331081338226795, 0.04336421564221382, -0.06963168829679489, -0.12501144409179688, 0.12125379592180252, 0.1339282989501953, -0.07799430936574936, 0.028010282665491104, -0.07278022915124893, 0.017418399453163147, -0.02815239131450653, -0.013141307048499584, -0.1986309140920639, -0.11172638088464737, 0.042703889310359955, -0.09334266930818558, 0.04449119046330452, -0.016933394595980644, 0.06300899386405945, 0.06127161905169487, -0.032935794442892075, -0.03907785564661026, -0.09548162668943405, -0.0007725379546172917, -0.10081403702497482, -0.192811518907547, -0.049560051411390305, -0.03276387229561806, 0.14925836026668549, -0.20759157836437225, 0.012256735935807228, 0.011097235605120659, 0.17082707583904266, 0.02680501900613308, -0.04922547563910484, 0.01144600659608841, 0.053548913449048996, -0.0179568100720644, -0.07918576896190643, 0.039481021463871, 0.006432307418435812, -0.13305583596229553, -0.03822280094027519, -0.11620089411735535, 0.03405029699206352, 0.09390159696340561, 0.031824808567762375, -0.08824271708726883, -0.045947980135679245, -0.0650644600391388, -0.0473501943051815, -0.050776220858097076, 0.000586897018365562, 0.19837966561317444, 0.02038322389125824, 0.13490264117717743, -0.061014432460069656, -0.06364881247282028, 0.0051582916639745235, 0.009017468430101871, -0.03921160101890564, 0.09328438341617584, 0.08292893320322037, -0.0862891897559166, 0.0845266655087471, 0.09053987264633179, -0.05654563382267952, 0.15276853740215302, -0.04727272689342499, -0.10035429149866104, -0.012882562354207039, 0.03481246158480644, -0.010597062297165394, 0.06594815105199814, -0.11019159108400345, 0.0013499516062438488, 0.029148871079087257, 0.026827991008758545, 0.053603142499923706, -0.16374120116233826, 0.015052767470479012, 0.023687897250056267, -0.03668319433927536, -0.0013466615928336978, -0.015691226348280907, 0.0244882944971323, 0.08318153768777847, 0.047949500381946564, -0.007533323019742966, 0.032784607261419296, -0.005198077764362097, -0.0767480656504631, 0.187256321310997, -0.13929930329322815, -0.17357687652111053, -0.13095450401306152, 0.03854134678840637, -0.07218790799379349, -0.03017667680978775, 0.025447634980082512, -0.10081613808870316, -0.05890035629272461, -0.0748816728591919, -0.009009233675897121, -0.07605826109647751, 0.01275542937219143, 0.0631902888417244, 0.014566747471690178, 0.06478860974311829, -0.13861681520938873, 0.014441946521401405, -0.015635671094059944, -0.07019011676311493, -0.0018146350048482418, 0.04804857820272446, 0.09328420460224152, 0.11665854603052139, -0.0014137272955849767, 0.01453313697129488, -0.03338881582021713, 0.2041899561882019, -0.07208313047885895, -0.02605801448225975, 0.09796129912137985, -0.009249108843505383, 0.06366020441055298, 0.08570584654808044, 0.044476468116045, -0.08391204476356506, 0.03376796469092369, 0.057778455317020416, -0.0031143033411353827, -0.261694997549057, -0.04745761305093765, -0.037771668285131454, -0.07067080587148666, 0.102826327085495, 0.06166237220168114, -0.020077327266335487, 0.05090496316552162, -0.02282474748790264, 0.0535476952791214, -0.024366576224565506, 0.08777303248643875, 0.10141105949878693, 0.04240220785140991, 0.08818431943655014, -0.03508289158344269, -0.03513142466545105, 0.07279076427221298, 0.014195103198289871, 0.2424984872341156, -0.015583381056785583, 0.13820409774780273, 0.034060172736644745, 0.12587159872055054, -0.011825131252408028, 0.030861666426062584, 0.02015690505504608, 0.010442309081554413, 0.0005363168311305344, -0.06430701166391373, -0.030793089419603348, 0.029687780886888504, 0.019189797341823578, 0.025199897587299347, -0.11046207696199417, 0.013323591090738773, 0.02245333045721054, 0.2656545042991638, 0.013024892657995224, -0.2906077206134796, -0.06843990832567215, 0.0142531031742692, -0.037062082439661026, -0.06545104086399078, 0.013440091162919998, 0.0989687442779541, -0.12735962867736816, 0.06870970875024796, -0.04992736130952835, 0.09173090755939484, -0.06575870513916016, -0.007395084016025066, 0.0294923335313797, 0.1476060152053833, -0.0023707293439656496, 0.0840994194149971, -0.21724161505699158, 0.19222961366176605, 0.018153060227632523, 0.10458042472600937, -0.06240173429250717, 0.04726595804095268, 0.020339706912636757, 0.03999387472867966, 0.08271463215351105, -0.010158051736652851, -0.06338363140821457, -0.1651887446641922, -0.0975838378071785, 0.037666160613298416, 0.14842039346694946, -0.022840389981865883, 0.0875270813703537, -0.03771791607141495, 0.007394794374704361, 0.04568314924836159, -0.07653490453958511, -0.18393373489379883, -0.1518574357032776, 0.03379305452108383, 0.036060940474271774, -0.00627728970721364, -0.07336213439702988, -0.11336588114500046, -0.03819717839360237, 0.19017934799194336, -0.02529752068221569, -0.049831751734018326, -0.1467987298965454, 0.0484158918261528, 0.13755933940410614, -0.06533658504486084, 0.02796148508787155, 0.00256528053432703, 0.14393970370292664, 0.026994140818715096, -0.09651829302310944, 0.06790591031312943, -0.06616368889808655, -0.1688041239976883, -0.04483600705862045, 0.12709392607212067, 0.05569593980908394, 0.03947494179010391, -0.010199020616710186, 0.014343684539198875, 0.0002922105777543038, -0.09214933961629868, 0.02390645444393158, 0.10033733397722244, 0.07402504980564117, 0.09494376182556152, -0.09386297315359116, 0.014087280258536339, -0.016184428706765175, -0.018560100346803665, 0.14881812036037445, 0.22151006758213043, -0.07118263095617294, 0.0939364805817604, 0.11817162483930588, -0.08216407150030136, -0.17385198175907135, 0.053498219698667526, 0.11901521682739258, 0.016335103660821915, 0.026910346001386642, -0.24145013093948364, 0.1026991605758667, 0.10991092771291733, -0.008935901336371899, 0.04382539913058281, -0.3207269310951233, -0.11998675018548965, 0.07621364295482635, 0.11226733028888702, 0.03775535151362419, -0.13262921571731567, -0.048193734139204025, -0.046045683324337006, -0.14899079501628876, 0.1283312439918518, -0.08815662562847137, 0.11971883475780487, -0.010755176655948162, 0.07683535665273666, 0.029183682054281235, -0.03730174899101257, 0.14385998249053955, 0.007379773072898388, 0.060625482350587845, -0.04678388684988022, 0.031108645722270012, 0.02299724519252777, -0.0657019093632698, 0.055765196681022644, -0.07960936427116394, 0.05870082601904869, -0.13202574849128723, -0.024444395676255226, -0.05211598053574562, 0.06935112923383713, -0.042769141495227814, -0.04929609224200249, -0.05191371589899063, 0.03183935582637787, 0.07449643313884735, -0.019971204921603203, 0.10725083947181702, 0.025936167687177658, 0.08392743021249771, 0.07984080910682678, 0.11276216804981232, -0.02200983464717865, -0.09951664507389069, -0.011613233014941216, -0.038146790117025375, 0.0753941461443901, -0.11311502754688263, 0.02082568034529686, 0.12511910498142242, 0.028173306956887245, 0.13569791615009308, 0.055305905640125275, -0.06266851723194122, 0.009064092300832272, 0.04968396946787834, -0.11187931895256042, -0.14057090878486633, -0.024362584576010704, -0.022774597629904747, -0.127961665391922, 0.023937921971082687, 0.11771246045827866, -0.04729795455932617, -0.013971985317766666, -0.010947117581963539, 0.01782352104783058, -0.0075651900842785835, 0.15619046986103058, 0.02935676835477352, 0.050170768052339554, -0.08457601070404053, 0.14339816570281982, 0.07385587692260742, -0.10591815412044525, 0.08327290415763855, 0.08608721196651459, -0.0769270732998848, -0.014690840616822243, 0.060490645468235016, 0.16178500652313232, -0.0338558591902256, -0.04966633394360542, -0.08370127528905869, -0.09706920385360718, 0.059599943459033966, 0.08319351077079773, 0.03671666979789734, 0.0017636590637266636, -0.03026546910405159, 0.048494528979063034, -0.14021289348602295, 0.07621302455663681, 0.045014236122369766, 0.07005853950977325, -0.13958393037319183, 0.114652119576931, 0.014361635781824589, 0.028740236535668373, -0.014928470365703106, 0.0002316843456355855, -0.11279858648777008, -0.02108139358460903, -0.14024648070335388, -0.005333472043275833, -0.03975614905357361, 0.010612674057483673, -0.008506640791893005, -0.035637594759464264, -0.045262373983860016, 0.04370827227830887, -0.07532915472984314, -0.06437861919403076, 0.007421700283885002, 0.07197430729866028, -0.13501471281051636, 0.00353173422627151, 0.018115028738975525, -0.10791123658418655, 0.0803743526339531, 0.05912632495164871, 0.013577395118772984, 0.03322415426373482, -0.12990179657936096, -0.024169020354747772, 0.03323298692703247, 0.021224552765488625, 0.04576818645000458, -0.08151385933160782, -0.009943020530045033, -0.022526120766997337, 0.04512377455830574, 0.0038537271320819855, 0.060443390160799026, -0.12495741993188858, -0.04681720212101936, -0.07381744682788849, -0.04721621051430702, -0.07885473221540451, 0.05740009620785713, 0.1124287098646164, 0.0384424589574337, 0.16026203334331512, -0.07709817588329315, 0.03464476764202118, -0.16868221759796143, -0.01978900283575058, -0.003193311160430312, -0.03217385709285736, -0.06685304641723633, -0.03399685397744179, 0.07805393636226654, -0.0640118196606636, 0.12794867157936096, -0.0469168983399868, 0.05884004384279251, 0.03164876624941826, -0.03447216749191284, -0.03823383152484894, 0.01902514509856701, 0.1999896615743637, 0.08348990231752396, -0.016639290377497673, 0.07042095810174942, -0.00965989287942648, 0.0710701048374176, 0.03754277154803276, 0.16804735362529755, 0.14130951464176178, -0.01194231677800417, 0.09371024370193481, 0.07324369996786118, -0.0950973629951477, -0.14862161874771118, 0.11402691155672073, -0.02415633201599121, 0.12601882219314575, -0.04355776682496071, 0.1796032339334488, 0.07824484258890152, -0.15247878432273865, 0.03852343559265137, -0.033963024616241455, -0.09628937393426895, -0.1144196093082428, -0.03592617064714432, -0.08702240139245987, -0.12760016322135925, 0.02134057879447937, -0.11968885362148285, 0.026334384456276894, 0.048974670469760895, 0.02151569165289402, 0.019481543451547623, 0.1277458220720291, -0.015691999346017838, -0.0012602005153894424, 0.0685315653681755, 0.005738239269703627, -0.006109815090894699, -0.055877409875392914, -0.09723249077796936, 0.05027436092495918, 0.013813922181725502, 0.07906186580657959, -0.029592664912343025, -0.012192969210445881, 0.046920645982027054, -0.014534810557961464, -0.07484496384859085, 0.029922135174274445, 0.014979821629822254, 0.033869896084070206, 0.048504140228033066, 0.04169824346899986, -0.02760489098727703, -0.050836991518735886, 0.24860879778862, -0.07370126247406006, -0.04401884227991104, -0.12845824658870697, 0.16312049329280853, 0.03323323652148247, -0.003308608429506421, 0.0670308917760849, -0.10790182650089264, -0.04682779684662819, 0.15938565135002136, 0.12899616360664368, -0.053771939128637314, -0.01850135251879692, 0.0059560565277934074, -0.01412162370979786, -0.04038148745894432, 0.12964218854904175, 0.1054653748869896, 0.04972127079963684, -0.043591003865003586, -0.014829757623374462, -0.008760049007833004, -0.03040069155395031, -0.09160114824771881, 0.059609346091747284, 0.013636612333357334, -0.01004059612751007, -0.04008077457547188, 0.06213417276740074, -0.024669766426086426, -0.16605189442634583, 0.013185592368245125, -0.14443686604499817, -0.16781985759735107, -0.03793634846806526, 0.0888485237956047, -0.010351672768592834, 0.05273124575614929, 0.00027353718178346753, -0.020218273624777794, 0.12218727171421051, -0.010056574828922749, -0.058384981006383896, -0.08303963392972946, 0.07272816449403763, -0.042770277708768845, 0.20786415040493011, 0.013068740256130695, 0.0756356492638588, 0.10810225456953049, 0.04395279660820961, -0.1257716566324234, 0.023292915895581245, 0.07003392279148102, -0.073834128677845, 0.04535720869898796, 0.14944639801979065, -0.057609569281339645, 0.0811193510890007, 0.03997379168868065, -0.10495609790086746, -0.009236905723810196, -0.05585689842700958, -0.027801835909485817, -0.06921965628862381, -0.005209196358919144, -0.06806754320859909, 0.14812447130680084, 0.20429904758930206, -0.031995031982660294, -0.018818458542227745, -0.06450250744819641, 0.03557588532567024, 0.04376384988427162, 0.08701568096876144, -0.026300817728042603, -0.21518835425376892, 0.008630834519863129, 0.041605621576309204, 0.04987727850675583, -0.23289278149604797, -0.09332188963890076, 0.027518123388290405, -0.03373486548662186, -0.07281313091516495, 0.10117164999246597, 0.07259813696146011, 0.027677245438098907, -0.04253042861819267, -0.13248272240161896, -0.0452192947268486, 0.1359308660030365, -0.1490495800971985, -0.04303785413503647 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert_bert_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "distilbert_bert_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/distilbert_bert_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# distilbert_bert_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
[ "# distilbert_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# distilbert_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ 62, 40, 6, 12, 8, 3, 118, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# distilbert_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ -0.08553527295589447, 0.1371486932039261, -0.003843798069283366, 0.07141382992267609, 0.13842305541038513, 0.038763340562582016, 0.10639432817697525, 0.11468085646629333, -0.10574610531330109, 0.05772063508629799, 0.07355295121669769, 0.06972567737102509, 0.04454325512051582, 0.15667371451854706, -0.03131961077451706, -0.2208627462387085, -0.000711853033863008, -0.007701047230511904, -0.08146040141582489, 0.11654456704854965, 0.08434736728668213, -0.10429486632347107, 0.05924871936440468, -0.0035208321642130613, -0.13368381559848785, 0.020300578325986862, -0.042379241436719894, -0.05015043914318085, 0.09517254680395126, -0.0036960355937480927, 0.09520036727190018, 0.03399626538157463, 0.14776340126991272, -0.20151019096374512, 0.003583107376471162, 0.10018743574619293, 0.04096946865320206, 0.09392526745796204, 0.060492321848869324, -0.01589386910200119, 0.13127455115318298, -0.15425018966197968, 0.10380198806524277, 0.018515927717089653, -0.07768109440803528, -0.11963590234518051, -0.10111874341964722, 0.05605239421129227, 0.0898517519235611, 0.11835552006959915, 0.006781754549592733, 0.1270504742860794, -0.10958874970674515, 0.08132586628198624, 0.17886459827423096, -0.20480279624462128, -0.047796398401260376, 0.05882064625620842, 0.023452138528227806, 0.0571809858083725, -0.09426911175251007, -0.017772672697901726, 0.023343084380030632, 0.022614169865846634, 0.10116158425807953, 0.0030459335539489985, -0.09437348693609238, -0.0020013204775750637, -0.1255306750535965, -0.01770375669002533, 0.07409191131591797, 0.02275715209543705, -0.009594086557626724, -0.09565937519073486, -0.05829542875289917, -0.14143843948841095, -0.024070506915450096, -0.012187778949737549, 0.038196861743927, -0.05636627599596977, -0.06581245362758636, -0.040188491344451904, -0.06462112814188004, -0.047149043530225754, -0.023238368332386017, 0.12246733903884888, 0.04556425288319588, 0.00725238723680377, -0.042683135718107224, 0.11486537009477615, 0.07445918023586273, -0.12431418895721436, 0.012637983076274395, 0.013426722027361393, -0.09465684741735458, -0.05046585947275162, -0.02585756778717041, -0.02178548276424408, -0.009902114979922771, 0.1347402036190033, -0.07361099123954773, 0.08140891045331955, 0.022888286039233208, -0.016537891700863838, -0.016606353223323822, 0.14741693437099457, -0.046751298010349274, -0.043245166540145874, -0.014255417510867119, 0.11084461212158203, 0.006140872370451689, -0.015034414827823639, -0.06831172108650208, 0.009576378390192986, 0.06381990015506744, 0.062347810715436935, -0.04099766165018082, 0.046533044427633286, -0.02034616470336914, -0.01806943118572235, 0.0464990958571434, -0.13762791454792023, 0.044637810438871384, 0.01985158585011959, -0.10218338668346405, -0.007636250462383032, 0.013939502649009228, -0.006585003342479467, -0.048330601304769516, 0.1313336044549942, -0.07659394294023514, 0.008322340436279774, -0.08261291682720184, -0.0804877057671547, 0.01352701522409916, -0.09819409996271133, -0.03896492347121239, -0.033955249935388565, -0.17346815764904022, -0.060089047998189926, 0.02965196967124939, -0.0608978196978569, -0.033987414091825485, -0.055844441056251526, -0.07413909584283829, 0.023637376725673676, -0.0075030019506812096, 0.17646966874599457, -0.06872162967920303, 0.06638070195913315, -0.024737896397709846, 0.030008193105459213, 0.06614281982183456, 0.047410476952791214, -0.07064270228147507, -0.008258980698883533, -0.0995357409119606, 0.08916875720024109, -0.08392324298620224, 0.006818512920290232, -0.10812335461378098, -0.09449239820241928, 0.02503957413136959, -0.009530584327876568, 0.055494412779808044, 0.13090971112251282, -0.20167860388755798, -0.0367959626019001, 0.11196187883615494, -0.05376549810171127, -0.007519596256315708, 0.05079098790884018, -0.056025635451078415, -0.009206375107169151, 0.06378649175167084, 0.12332527339458466, 0.06868711858987808, -0.12382917106151581, -0.007068349048495293, 0.0029158806428313255, 0.037439245730638504, 0.013223806396126747, 0.015386797487735748, 0.004952809773385525, 0.06500450521707535, 0.006755456328392029, -0.0391293428838253, 0.009401801973581314, -0.0958646610379219, -0.06554246693849564, -0.03379368036985397, -0.07397275418043137, 0.021540168672800064, 0.031145034357905388, 0.03197188302874565, -0.056934602558612823, -0.1170707643032074, 0.08829493820667267, 0.12335050106048584, -0.05674939602613449, 0.015230138786137104, -0.07512405514717102, -0.03606751933693886, 0.01646312139928341, -0.01966742053627968, -0.19894394278526306, -0.11764129251241684, 0.021346287801861763, -0.06467211991548538, 0.04216393455862999, -0.010381543077528477, 0.06525079160928726, 0.04795410856604576, -0.03784044831991196, -0.0185215026140213, -0.061155762523412704, 0.004631508141756058, -0.09045787900686264, -0.21300561726093292, -0.03508496284484863, -0.016810307279229164, 0.15867777168750763, -0.21950244903564453, 0.01239447295665741, -0.016511211171746254, 0.14786843955516815, 0.024679025635123253, -0.05784298852086067, 0.0120172630995512, 0.05256407707929611, 0.006376360543072224, -0.09567886590957642, 0.032769620418548584, -0.010538860224187374, -0.07632946968078613, -0.01670418679714203, -0.13932418823242188, -0.024162495508790016, 0.08050481975078583, 0.08751093596220016, -0.10388610512018204, 0.021240556612610817, -0.07430337369441986, -0.04110459238290787, -0.08969586342573166, 0.03720585256814957, 0.20479871332645416, 0.03642631322145462, 0.1306391805410385, -0.045101676136255264, -0.0848192423582077, -0.01065538078546524, 0.02319452166557312, 0.011347874999046326, 0.1080610603094101, 0.08231890946626663, -0.04343204200267792, 0.07403769344091415, 0.022532736882567406, -0.039433687925338745, 0.13050417602062225, -0.04571248218417168, -0.08482448756694794, 0.0023699717130512, -0.029805677011609077, -0.023921053856611252, 0.0987817645072937, -0.03395642340183258, 0.005388610530644655, 0.03181208670139313, 0.024581344798207283, 0.03175465017557144, -0.18512694537639618, 0.002383090788498521, 0.016690554097294807, -0.05634577199816704, -0.03898888826370239, -0.021060699597001076, 0.05484146624803543, 0.10211235284805298, 0.0031813494861125946, -0.0334957018494606, 0.009311620146036148, -0.007148273754864931, -0.0686316192150116, 0.17578303813934326, -0.11656476557254791, -0.16775959730148315, -0.07808500528335571, 0.018541786819696426, -0.030275685712695122, -0.04977671056985855, 0.003006989834830165, -0.11775153130292892, -0.07314109802246094, -0.11291797459125519, -0.013901172205805779, 0.010597464628517628, -0.0009591180714778602, 0.04391435533761978, 0.01258462481200695, 0.052099429070949554, -0.14026664197444916, 0.014391240663826466, -0.05094563961029053, -0.07763606309890747, 0.017865777015686035, 0.0814804658293724, 0.06054118648171425, 0.15245774388313293, -0.013842454180121422, 0.02797112613916397, -0.01604374870657921, 0.1963237226009369, -0.09210806339979172, 0.015991169959306717, 0.10068297386169434, 0.016319269314408302, 0.04386289417743683, 0.12323909252882004, 0.0419333279132843, -0.0772407129406929, 0.02777881920337677, 0.09899834543466568, -0.02208961546421051, -0.25617703795433044, -0.0664602741599083, -0.016415884718298912, -0.08377955108880997, 0.08645226061344147, 0.05535460636019707, -0.011459228582680225, 0.004285136703401804, -0.009353232569992542, -0.014013116247951984, -0.0024128123186528683, 0.04984021186828613, 0.08089537918567657, 0.054484933614730835, 0.09390096366405487, -0.03128822520375252, -0.025810493156313896, 0.06310941278934479, -0.008719164878129959, 0.22405706346035004, -0.06570947170257568, 0.074248768389225, 0.021268628537654877, 0.08820130676031113, -0.014049317687749863, 0.02908330410718918, 0.023504476994276047, -0.013619978912174702, 0.013722559437155724, -0.0483878031373024, 0.0003870640939567238, 0.0001087808923330158, -0.013455337844789028, -0.003518456593155861, -0.06688972562551498, 0.037783849984407425, 0.015815965831279755, 0.273346871137619, 0.03835581615567207, -0.2867255210876465, -0.052316296845674515, -0.022956514731049538, -0.03162246197462082, -0.04479886218905449, -0.009909248910844326, 0.09204208850860596, -0.11195682734251022, 0.06524567306041718, -0.04396824911236763, 0.08393017947673798, -0.07582291215658188, -0.001043393393047154, 0.06961172819137573, 0.14608201384544373, -0.01621900126338005, 0.059104204177856445, -0.2140766978263855, 0.2198847383260727, 0.007536973338574171, 0.12297677993774414, -0.06492206454277039, 0.021055473014712334, 0.016491372138261795, 0.04472854733467102, 0.061050187796354294, -0.004502039402723312, -0.02577180415391922, -0.15990395843982697, -0.11641133576631546, 0.022848524153232574, 0.11592739820480347, -0.0006815774831920862, 0.0816926434636116, -0.025422710925340652, -0.018760820850729942, 0.04073518514633179, -0.15940029919147491, -0.1554400622844696, -0.1266312599182129, 0.033908165991306305, 0.02826056256890297, -0.04010023549199104, -0.056607235223054886, -0.11970622092485428, -0.021320784464478493, 0.17217449843883514, 0.023725828155875206, -0.062052804976701736, -0.14721795916557312, 0.06334247440099716, 0.15129901468753815, -0.04391952604055405, 0.010147872380912304, 0.03567162901163101, 0.11796160042285919, 0.047684356570243835, -0.08731675148010254, 0.03762783482670784, -0.06397613137960434, -0.18817487359046936, -0.05703962966799736, 0.13113731145858765, 0.09201125800609589, 0.04984108731150627, -0.02297487109899521, 0.028181571513414383, 0.004793635103851557, -0.07985085248947144, 0.006213079672306776, 0.08714782446622849, 0.07279302924871445, 0.05718438699841499, -0.06800362467765808, -0.00023277808213606477, -0.04388364776968956, -0.0188997071236372, 0.1008550301194191, 0.18947604298591614, -0.07902564108371735, 0.11054090410470963, 0.06576371937990189, -0.06841935962438583, -0.17854101955890656, 0.05664096027612686, 0.1272478699684143, 0.008346864022314548, 0.03167816996574402, -0.2043418139219284, 0.12496434152126312, 0.11782238632440567, -0.01010619942098856, 0.05291200801730156, -0.3534694314002991, -0.12645626068115234, 0.05643874406814575, 0.0826077088713646, 0.01591579243540764, -0.10413624346256256, -0.02172805555164814, -0.016027875244617462, -0.133791983127594, 0.1422751247882843, -0.05726764723658562, 0.10936269164085388, -0.006260668858885765, 0.10225321352481842, 0.023387059569358826, -0.0421268567442894, 0.10552804172039032, 0.06392977386713028, 0.06992232799530029, -0.04635380953550339, 0.018301082774996758, 0.03582325950264931, -0.06482510268688202, 0.03445841744542122, -0.04628061130642891, 0.06893961131572723, -0.11634483933448792, -0.01383913867175579, -0.08405635505914688, 0.04607338458299637, -0.04504961520433426, -0.044366635382175446, -0.03870631381869316, 0.052605073899030685, 0.07779810577630997, -0.0475715771317482, 0.06454074382781982, 0.005894320085644722, 0.083851657807827, 0.06508471071720123, 0.08584009855985641, -0.05656803026795387, -0.09739188104867935, -0.007683895528316498, -0.007896347902715206, 0.033157337456941605, -0.11343733221292496, 0.025737425312399864, 0.1364835947751999, 0.0497158020734787, 0.12194254249334335, 0.03571327403187752, -0.036633122712373734, -0.020562438294291496, 0.035347312688827515, -0.11945122480392456, -0.09160173684358597, 0.030812762677669525, -0.06531069427728653, -0.10469279438257217, 0.014114352874457836, 0.10282028466463089, -0.03397299349308014, -0.01885705441236496, 0.003600588534027338, 0.03114621341228485, 0.015830527991056442, 0.2185710370540619, 0.027156945317983627, 0.06552302837371826, -0.11618860065937042, 0.1258959174156189, 0.06334754079580307, -0.0901370570063591, 0.045037299394607544, 0.1303216814994812, -0.1009758710861206, -0.01253057923167944, 0.10037557780742645, 0.1366065889596939, -0.03261777386069298, -0.027896253392100334, -0.09224054962396622, -0.09592133015394211, 0.07931897789239883, 0.13128697872161865, 0.03113788180053234, 0.0028559130150824785, -0.05559471622109413, 0.014666307717561722, -0.14990442991256714, 0.06768466532230377, 0.08486512303352356, 0.05489855259656906, -0.0844779759645462, 0.13784107565879822, 0.027204804122447968, 0.031021634116768837, -0.017781954258680344, 0.014443722553551197, -0.07378067076206207, -0.012115978635847569, -0.0884949341416359, -0.014897085726261139, -0.01012603659182787, -0.0019321443978697062, -0.022146310657262802, -0.050019990652799606, -0.037709545344114304, 0.04507941007614136, -0.07613541930913925, -0.05995292216539383, -0.002726600505411625, 0.03992339223623276, -0.14126501977443695, 0.0024363293778151274, 0.004282911773771048, -0.09404876828193665, 0.06923723965883255, 0.04538392275571823, -0.0045740557834506035, 0.03368972986936569, -0.1555134505033493, -0.04595661908388138, 0.030684493482112885, 0.01674313098192215, 0.08187270909547806, -0.07736345380544662, -0.0025582651142030954, -0.019990630447864532, 0.0385282076895237, 0.013695130124688148, 0.04940248280763626, -0.11297149956226349, 0.007272363640367985, -0.07030193507671356, -0.034431006759405136, -0.06891932338476181, 0.04286740720272064, 0.11016900837421417, 0.042847223579883575, 0.17735555768013, -0.08114150911569595, 0.04046071320772171, -0.18106701970100403, -0.032541532069444656, -0.0035857923794537783, -0.04069133847951889, -0.06807807832956314, -0.014991267584264278, 0.11058079451322556, -0.053053420037031174, 0.12192948162555695, 0.013414107263088226, 0.07188001275062561, 0.041860975325107574, -0.02903607115149498, -0.06393850594758987, 0.015023328363895416, 0.1418241262435913, 0.06085359677672386, -0.019770624116063118, 0.11325953155755997, -0.011152729392051697, 0.05059748888015747, 0.05000954866409302, 0.22892002761363983, 0.13022494316101074, 0.02551661804318428, 0.06794323772192001, 0.047522637993097305, -0.14403089880943298, -0.12017340213060379, 0.10668205469846725, -0.09347163140773773, 0.12736956775188446, -0.06694581359624863, 0.21080298721790314, 0.048440996557474136, -0.14472484588623047, 0.05516316369175911, -0.030677763745188713, -0.10427005589008331, -0.1047043427824974, -0.00870518758893013, -0.07698824256658554, -0.11062624305486679, 0.034999050199985504, -0.11031399667263031, 0.06664206832647324, 0.09385992586612701, 0.030586296692490578, 0.02708020992577076, 0.11321348696947098, -0.0048998561687767506, 0.000007249716873047873, 0.05775507539510727, 0.025785323232412338, -0.00599126610904932, -0.06516290456056595, -0.06416191160678864, 0.02328403666615486, 0.02546437457203865, 0.08076130598783493, -0.03568841889500618, -0.008787238970398903, 0.044464319944381714, -0.01766274683177471, -0.07570458948612213, 0.03913883492350578, 0.0013494376325979829, 0.0657137930393219, 0.05820370838046074, 0.055912893265485764, 0.006995458155870438, -0.048361942172050476, 0.2867644429206848, -0.06283219903707504, -0.1112573966383934, -0.12306907773017883, 0.23802721500396729, 0.057863906025886536, -0.022099614143371582, 0.0633346363902092, -0.1051260232925415, -0.04862665385007858, 0.1553044617176056, 0.17037850618362427, -0.05901998281478882, -0.026859885081648827, -0.03457273915410042, -0.01227694284170866, -0.029525121673941612, 0.12717579305171967, 0.1118919774889946, 0.05630898475646973, -0.0472438782453537, -0.023084623739123344, -0.018924064934253693, -0.042335715144872665, -0.07291916012763977, 0.07542381435632706, 0.021871965378522873, -0.0036319901701062918, -0.036510273814201355, 0.069781593978405, -0.010882865637540817, -0.21184758841991425, 0.03976447880268097, -0.16913636028766632, -0.1850544661283493, -0.00879126600921154, 0.09819858521223068, -0.022580867633223534, 0.04912091791629791, 0.01808888278901577, -0.015266628004610538, 0.10830379277467728, -0.02945186011493206, -0.03494875133037567, -0.10500723868608475, 0.0645008459687233, -0.1306823194026947, 0.22972333431243896, -0.00732895964756608, 0.06620761752128601, 0.09286017715930939, 0.02518334612250328, -0.12647739052772522, 0.06291088461875916, 0.04693542793393135, -0.09971025586128235, 0.03121240809559822, 0.16116118431091309, -0.04240468144416809, 0.045508451759815216, 0.03808083385229111, -0.12217721343040466, -0.0013317574048414826, -0.08908805251121521, -0.03445812314748764, -0.06145206838846207, -0.01815861277282238, -0.02812943235039711, 0.14726023375988007, 0.21814607083797455, -0.008099148981273174, 0.028888512402772903, -0.08047805726528168, -0.006140727549791336, 0.023156948387622833, 0.08234650641679764, -0.028904521837830544, -0.2236606478691101, 0.03418426960706711, 0.009370905347168446, 0.03825995326042175, -0.1765059530735016, -0.0725872740149498, 0.0059500704519450665, -0.0704878643155098, -0.051037270575761795, 0.11079076677560806, 0.04419872537255287, 0.04101628437638283, -0.024602368474006653, -0.07761659473180771, -0.010818042792379856, 0.1481967568397522, -0.19039268791675568, -0.03466469421982765 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert_distilgpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "distilbert_distilgpt2_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/distilbert_distilgpt2_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# distilbert_distilgpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
[ "# distilbert_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# distilbert_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ 62, 44, 6, 12, 8, 3, 118, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# distilbert_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ -0.085886649787426, 0.15049423277378082, -0.0036458149552345276, 0.07377687096595764, 0.1286974847316742, 0.041172657161951065, 0.11018669605255127, 0.13400407135486603, -0.10262317210435867, 0.0651286318898201, 0.08305195719003677, 0.044070858508348465, 0.053792838007211685, 0.15862520039081573, -0.030705230310559273, -0.23130467534065247, 0.007431476376950741, -0.009299867786467075, -0.06200410798192024, 0.11895682662725449, 0.08847769349813461, -0.09305267035961151, 0.06247302144765854, -0.01113489456474781, -0.12948748469352722, 0.018175333738327026, -0.04198634624481201, -0.05838508531451225, 0.08978509157896042, -0.017030077055096626, 0.09200303256511688, 0.032845109701156616, 0.14385266602039337, -0.20586036145687103, 0.0000019601297935878392, 0.08370915055274963, 0.03712012991309166, 0.08808606117963791, 0.06291438639163971, -0.020649515092372894, 0.11248351633548737, -0.17793577909469604, 0.09880230575799942, 0.014054051600396633, -0.08054392784833908, -0.09315051138401031, -0.10079336911439896, 0.0762118324637413, 0.097889743745327, 0.11217121034860611, 0.005878110881894827, 0.12009156495332718, -0.10363546758890152, 0.08463867008686066, 0.1755979061126709, -0.21314015984535217, -0.05656927451491356, 0.06122670695185661, 0.020855970680713654, 0.06137464568018913, -0.0912439376115799, -0.01585937663912773, 0.029025334864854813, 0.024067070335149765, 0.0995086282491684, 0.0009277049684897065, -0.0943317636847496, 0.002019183710217476, -0.12600597739219666, -0.041368380188941956, 0.11956947296857834, 0.02152775228023529, -0.016791507601737976, -0.1072617769241333, -0.04837483912706375, -0.1414117068052292, -0.013734117150306702, -0.013398027047514915, 0.035714250057935715, -0.061475593596696854, -0.05821843445301056, -0.05502352863550186, -0.07739148288965225, -0.047705311328172684, -0.009720929898321629, 0.08646587282419205, 0.04152122512459755, 0.012902902439236641, -0.0427057109773159, 0.12476405501365662, 0.06427143514156342, -0.12757955491542816, -0.0030549399089068174, 0.011448152363300323, -0.08865701407194138, -0.05531753599643707, -0.01906740851700306, -0.021886905655264854, -0.010422434657812119, 0.1359676718711853, -0.058282654732465744, 0.07553750276565552, 0.020552081987261772, -0.01149903703480959, -0.018333453685045242, 0.13732965290546417, -0.03457753732800484, -0.04828321561217308, -0.016032855957746506, 0.11623048782348633, 0.0032213996164500713, -0.009509188123047352, -0.06768163293600082, -0.004667313303798437, 0.07843297719955444, 0.06536444276571274, -0.031342122703790665, 0.04311935603618622, -0.01899721845984459, -0.0228040162473917, 0.036767300218343735, -0.1466161012649536, 0.03898245096206665, 0.011132022365927696, -0.09955235570669174, -0.0012227896368131042, 0.024610593914985657, -0.012716324999928474, -0.05511903762817383, 0.11639834940433502, -0.07327821850776672, 0.006539834663271904, -0.07611549645662308, -0.06706851720809937, 0.023991120979189873, -0.109119713306427, -0.04712885245680809, -0.03337147459387779, -0.2062941938638687, -0.06396941840648651, 0.020241931080818176, -0.054619286209344864, -0.04760974273085594, -0.06393665820360184, -0.07933998852968216, 0.019174784421920776, -0.0073556010611355305, 0.15960495173931122, -0.058926258236169815, 0.07001851499080658, -0.023608041927218437, 0.026486780494451523, 0.07371550798416138, 0.043297071009874344, -0.06708332896232605, 0.005472090095281601, -0.10019369423389435, 0.09307621419429779, -0.0852913036942482, 0.007985572330653667, -0.10885199904441833, -0.10215678066015244, 0.01720450446009636, -0.017563579604029655, 0.041205327957868576, 0.14965350925922394, -0.19274850189685822, -0.030765604227781296, 0.13753244280815125, -0.06438720226287842, -0.007764481473714113, 0.06262096762657166, -0.05431954935193062, -0.0005443756235763431, 0.06523022055625916, 0.13473421335220337, 0.08099785447120667, -0.12327756732702255, -0.019334860146045685, -0.002114813541993499, 0.03851094841957092, 0.011965708807110786, 0.029611919075250626, 0.007658923044800758, 0.0665440782904625, 0.010774285532534122, -0.06958656758069992, -0.003622656222432852, -0.09175858646631241, -0.0727955624461174, -0.04251932352781296, -0.07591186463832855, 0.0394652895629406, 0.023495400324463844, 0.04039645195007324, -0.05950303375720978, -0.10691279917955399, 0.09390148520469666, 0.12628337740898132, -0.05564699321985245, 0.007198739796876907, -0.0766419917345047, -0.03018866293132305, 0.013765053823590279, -0.02293628640472889, -0.1942218542098999, -0.12497062236070633, 0.028972387313842773, -0.05019056051969528, 0.04537767544388771, -0.00848220195621252, 0.06612195819616318, 0.03831668198108673, -0.040274567902088165, -0.02738153003156185, -0.06772205233573914, -0.0037775032687932253, -0.08655598014593124, -0.18690125644207, -0.048463668674230576, -0.017697511240839958, 0.1707860827445984, -0.22462452948093414, 0.029547475278377533, 0.0007644413271918893, 0.1503816694021225, 0.02331845834851265, -0.06287109851837158, 0.02478140965104103, 0.0506100133061409, 0.009271878749132156, -0.09503274410963058, 0.02662193402647972, -0.01098676212131977, -0.0711933970451355, -0.025543691590428352, -0.13150011003017426, -0.020333167165517807, 0.0680411159992218, 0.10376934707164764, -0.1120632216334343, 0.013724750839173794, -0.07185236364603043, -0.05105387419462204, -0.10205050557851791, 0.01318433228880167, 0.19623373448848724, 0.036767780780792236, 0.1272924691438675, -0.04608311504125595, -0.08637198060750961, -0.006967164110392332, 0.017401712015271187, 0.001006276928819716, 0.10007837414741516, 0.05812031775712967, -0.05685582011938095, 0.08272840082645416, 0.014997235499322414, -0.04725867882370949, 0.1429813802242279, -0.05475110933184624, -0.08989375084638596, -0.008913883939385414, -0.00747038796544075, -0.02065379172563553, 0.09912095218896866, -0.04102182388305664, -0.015122050419449806, 0.03321785479784012, 0.02069583162665367, 0.03176867961883545, -0.17702637612819672, -0.0051873126067221165, 0.016644209623336792, -0.06596474349498749, -0.024143008515238762, -0.0218056533485651, 0.054404743015766144, 0.09767360240221024, -0.00013200778630562127, -0.02730894833803177, 0.018350956961512566, -0.011638563126325607, -0.08197134733200073, 0.17129972577095032, -0.11099493503570557, -0.1746612936258316, -0.10732757300138474, 0.050535961985588074, -0.03540080785751343, -0.05127958208322525, 0.002090707654133439, -0.10095009207725525, -0.06031502038240433, -0.1132001280784607, -0.032562777400016785, -0.004580584354698658, -0.009986615739762783, 0.04175662249326706, 0.017043817788362503, 0.051333751529455185, -0.13124366104602814, 0.012598126195371151, -0.02935011312365532, -0.08109451085329056, 0.024247342720627785, 0.05810733884572983, 0.06439218670129776, 0.1393147110939026, -0.014246542006731033, 0.031846191734075546, -0.022892428562045097, 0.19243448972702026, -0.09171772748231888, 0.0194456335157156, 0.10972438752651215, 0.01310573611408472, 0.05105937644839287, 0.10769712924957275, 0.0315287709236145, -0.0646698847413063, 0.020811237394809723, 0.08301057666540146, -0.023684943094849586, -0.26439401507377625, -0.06012710928916931, -0.014860082417726517, -0.055267930030822754, 0.0971532091498375, 0.0615694597363472, 0.007406558841466904, 0.014229544438421726, -0.014902612194418907, -0.01226570550352335, -0.002959523582831025, 0.061883989721536636, 0.08566964417695999, 0.043706074357032776, 0.08853000402450562, -0.0333501473069191, -0.027630779892206192, 0.06176256388425827, 0.004361435305327177, 0.2366645187139511, -0.05115335434675217, 0.09215670824050903, 0.012003939598798752, 0.10726085305213928, -0.01386367715895176, 0.03669220581650734, 0.03274795040488243, 0.0011781163047999144, 0.02023213729262352, -0.049501508474349976, -0.005400907248258591, 0.011740508489310741, -0.007734612561762333, 0.0014605809701606631, -0.06613215804100037, 0.03461570292711258, 0.015062982216477394, 0.3027949929237366, 0.04227583482861519, -0.2891767919063568, -0.055564288049936295, -0.012159241363406181, -0.04371064528822899, -0.047619763761758804, -0.013231257908046246, 0.10833761096000671, -0.1380712389945984, 0.06938934326171875, -0.04568571224808693, 0.0847618579864502, -0.07795412093400955, -0.0032698737923055887, 0.05473783612251282, 0.10782992094755173, -0.012218677438795567, 0.06778624653816223, -0.20405398309230804, 0.22301575541496277, 0.00993357039988041, 0.11540277302265167, -0.07141413539648056, 0.024776583537459373, 0.018084365874528885, 0.04549353942275047, 0.08986788243055344, -0.003907396458089352, -0.037148915231227875, -0.1516685038805008, -0.13562443852424622, 0.016230562701821327, 0.11515011638402939, -0.026758568361401558, 0.0856083407998085, -0.016822565346956253, -0.017360132187604904, 0.02601092867553234, -0.1515512317419052, -0.14646507799625397, -0.12116899341344833, 0.03965218365192413, 0.031552281230688095, -0.02658272720873356, -0.05979955941438675, -0.11690264195203781, -0.01936882920563221, 0.18303237855434418, -0.01074936892837286, -0.06741335242986679, -0.14597690105438232, 0.05680307745933533, 0.1474488079547882, -0.043628666549921036, 0.008929899893701077, 0.04244012013077736, 0.12164904922246933, 0.03770610690116882, -0.08175650238990784, 0.016428906470537186, -0.0597122386097908, -0.18823407590389252, -0.04871411249041557, 0.14371055364608765, 0.08215969800949097, 0.05204810947179794, -0.015133308246731758, 0.02764696627855301, 0.011145873926579952, -0.07836311310529709, 0.005878072697669268, 0.09006714075803757, 0.08886346966028214, 0.06626425683498383, -0.07035955786705017, -0.0038527799770236015, -0.053945161402225494, -0.019821424037218094, 0.11040576547384262, 0.17021098732948303, -0.08274304121732712, 0.10626067966222763, 0.04575352743268013, -0.07373111695051193, -0.1700952649116516, 0.0482812263071537, 0.13354673981666565, 0.019400298595428467, 0.0395674966275692, -0.20362813770771027, 0.1085255965590477, 0.12670733034610748, -0.013691955246031284, 0.04539181664586067, -0.34954506158828735, -0.12082835286855698, 0.05040816217660904, 0.0677700862288475, 0.005868744105100632, -0.11383114755153656, -0.03343476355075836, 0.004949682392179966, -0.12856511771678925, 0.13404807448387146, -0.05420632287859917, 0.10346908122301102, -0.00975390337407589, 0.12139250338077545, 0.027249345555901527, -0.04673399776220322, 0.12354782223701477, 0.08170797675848007, 0.06278335303068161, -0.0472467802464962, 0.003968316596001387, 0.057833198457956314, -0.07945100218057632, 0.0696134865283966, -0.04492412880063057, 0.06572762876749039, -0.1557973176240921, -0.012245671823620796, -0.08447791635990143, 0.04767070710659027, -0.04834533855319023, -0.03655446693301201, -0.031395670026540756, 0.054771989583969116, 0.06855489313602448, -0.03957512974739075, 0.06044560670852661, 0.007832263596355915, 0.08626137673854828, 0.09304217994213104, 0.09418615698814392, -0.017423095181584358, -0.11468592286109924, -0.019145645201206207, -0.012148148380219936, 0.031189726665616035, -0.1123402938246727, 0.013268468901515007, 0.13889703154563904, 0.05691073089838028, 0.1360313892364502, 0.026516174897551537, -0.042658936232328415, -0.017876820638775826, 0.03459731116890907, -0.12934215366840363, -0.0988394245505333, 0.01165973674505949, -0.05916581302881241, -0.12437567114830017, 0.005138395819813013, 0.10558564960956573, -0.04059041664004326, -0.022719964385032654, 0.0004248463665135205, 0.03367885202169418, 0.010225930251181126, 0.21164490282535553, 0.025675375014543533, 0.07441936433315277, -0.10616781562566757, 0.11395271122455597, 0.0710965245962143, -0.09584121406078339, 0.048135168850421906, 0.13136987388134003, -0.09615062922239304, -0.015884235501289368, 0.08750265836715698, 0.1337164044380188, -0.033336274325847626, -0.030968165025115013, -0.08714226633310318, -0.09441910684108734, 0.07475525885820389, 0.1279415488243103, 0.036243464797735214, 0.015996713191270828, -0.04702233523130417, 0.008373984135687351, -0.13735052943229675, 0.07059367746114731, 0.09713021665811539, 0.05709181725978851, -0.0961364209651947, 0.15087422728538513, 0.018760355189442635, 0.04444293677806854, -0.013061697594821453, 0.010253922082483768, -0.06562123447656631, -0.01125183142721653, -0.0711052194237709, -0.0013375496491789818, -0.018409475684165955, -0.006200718693435192, -0.015527725219726562, -0.044984664767980576, -0.032235000282526016, 0.04060325771570206, -0.07323286682367325, -0.06028086692094803, -0.007369858678430319, 0.049750108271837234, -0.13444262742996216, -0.0016435075085610151, 0.007918051443994045, -0.10036716610193253, 0.07017930597066879, 0.05614457279443741, -0.003297031158581376, 0.025269202888011932, -0.10850636661052704, -0.02634579874575138, 0.028514308854937553, 0.02011040598154068, 0.06572957336902618, -0.07237276434898376, -0.0021289053838700056, -0.02397734671831131, 0.02844763547182083, 0.028089074417948723, 0.0410301499068737, -0.11346820741891861, 0.00939969439059496, -0.060696277767419815, -0.030428243800997734, -0.07586918026208878, 0.04398167133331299, 0.11832566559314728, 0.03493054211139679, 0.1668694168329239, -0.08179300278425217, 0.04931430146098137, -0.18791444599628448, -0.028742434456944466, -0.0023314887657761574, -0.03927856683731079, -0.05459737777709961, -0.014594940468668938, 0.10769324004650116, -0.05046303570270538, 0.10798940062522888, 0.004687268752604723, 0.0621880441904068, 0.03942498937249184, -0.03238595649600029, -0.062273431569337845, 0.014373699203133583, 0.13236020505428314, 0.0582004114985466, -0.024036971852183342, 0.10877490788698196, -0.026800772175192833, 0.03357598930597305, 0.036303773522377014, 0.218819260597229, 0.12631084024906158, 0.006529642269015312, 0.07185850292444229, 0.06000666320323944, -0.12596441805362701, -0.1258729100227356, 0.09669487178325653, -0.07407663762569427, 0.1233837902545929, -0.06465316563844681, 0.1917162835597992, 0.06254272907972336, -0.15918132662773132, 0.057789966464042664, -0.025836456567049026, -0.10890607535839081, -0.0998595654964447, -0.04053314030170441, -0.07478759437799454, -0.09290506690740585, 0.031964175403118134, -0.1174081340432167, 0.06232643127441406, 0.08950500190258026, 0.0300289373844862, 0.019272591918706894, 0.11574170738458633, -0.01381215825676918, -0.0011494426289573312, 0.07757903635501862, 0.020263392478227615, 0.003444465808570385, -0.07414597272872925, -0.0619838610291481, 0.028450530022382736, 0.030762730166316032, 0.08469411730766296, -0.034493256360292435, 0.014463555999100208, 0.029714157804846764, -0.020176194608211517, -0.0754416286945343, 0.03235882148146629, 0.0029295755084604025, 0.06314082443714142, 0.06239647790789604, 0.05861208587884903, -0.0027758204378187656, -0.04974082484841347, 0.27062055468559265, -0.06569726765155792, -0.10527937859296799, -0.13636378943920135, 0.210431307554245, 0.043338898569345474, -0.03010503575205803, 0.06804420799016953, -0.10722710192203522, -0.031550660729408264, 0.154274582862854, 0.16599902510643005, -0.03517702594399452, -0.01792583055794239, -0.035634011030197144, -0.010052905417978764, -0.012539215385913849, 0.10400723665952682, 0.08932006359100342, 0.06701869517564774, -0.05253727734088898, -0.014724037609994411, -0.009192223660647869, -0.04016866534948349, -0.08840472996234894, 0.07046455144882202, 0.013133738189935684, -0.0020593972876667976, -0.03454456105828285, 0.08115433901548386, -0.01911558024585247, -0.18680691719055176, 0.04241788014769554, -0.17660506069660187, -0.18481002748012543, -0.0047149211168289185, 0.08796815574169159, -0.02294793911278248, 0.046208836138248444, 0.013154787011444569, -0.017969613894820213, 0.11338387429714203, -0.02997639775276184, -0.05177351459860802, -0.09821628779172897, 0.07167854905128479, -0.11212114244699478, 0.2173445224761963, -0.0009239425417035818, 0.07932637631893158, 0.08454020321369171, 0.014321336522698402, -0.13123153150081635, 0.047182902693748474, 0.047943275421857834, -0.07173538953065872, 0.02610664814710617, 0.17967994511127472, -0.04219166934490204, 0.048507895320653915, 0.03225291892886162, -0.12333782017230988, -0.016851797699928284, -0.04810749739408493, -0.014224804006516933, -0.07302361726760864, -0.012210499495267868, -0.026240479201078415, 0.15257304906845093, 0.2174014449119568, -0.01683618873357773, 0.022257942706346512, -0.07071854919195175, -0.005668330006301403, 0.029331335797905922, 0.07628417760133743, -0.021155834197998047, -0.21571892499923706, 0.02362598292529583, -0.019828209653496742, 0.04035266861319542, -0.18772557377815247, -0.08514413982629776, 0.012873181141912937, -0.06002268195152283, -0.05086100846529007, 0.10798706859350204, 0.052866362035274506, 0.03601615130901337, -0.01624876819550991, -0.05916345492005348, -0.012970194220542908, 0.14732535183429718, -0.18760062754154205, -0.03684154525399208 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert_gpt2_summarization_cnndm This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "distilbert_gpt2_summarization_cnndm", "results": []}]}
text2text-generation
Ayham/distilbert_gpt2_summarization_cnndm
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# distilbert_gpt2_summarization_cnndm This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# distilbert_gpt2_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# distilbert_gpt2_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 62, 40, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# distilbert_gpt2_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.08500642329454422, 0.13782647252082825, -0.003898640163242817, 0.0692664310336113, 0.1419195830821991, 0.041762929409742355, 0.0951128751039505, 0.11522649973630905, -0.11586043983697891, 0.05759895592927933, 0.07737758755683899, 0.07106632739305496, 0.047620609402656555, 0.15088526904582977, -0.03332209214568138, -0.22394530475139618, 0.003500190796330571, -0.01014753058552742, -0.08763686567544937, 0.11454853415489197, 0.08654426783323288, -0.10516583174467087, 0.05866168811917305, -0.0016005398938432336, -0.1407882124185562, 0.017485838383436203, -0.04489412158727646, -0.049736153334379196, 0.09346260875463486, -0.004441660828888416, 0.09759727865457535, 0.03652448579668999, 0.14840617775917053, -0.20536115765571594, 0.002509061712771654, 0.09653417021036148, 0.04626960679888725, 0.09651308506727219, 0.05795327574014664, -0.007786651141941547, 0.1255093365907669, -0.1519274115562439, 0.09911974519491196, 0.01773889735341072, -0.07462036609649658, -0.12111802399158478, -0.0957353338599205, 0.05462956055998802, 0.08723541349172592, 0.11523859202861786, 0.006295831874012947, 0.12017780542373657, -0.10819999873638153, 0.07909092307090759, 0.16671395301818848, -0.20992860198020935, -0.052042748779058456, 0.06293178349733353, 0.030399886891245842, 0.05697319656610489, -0.09461094439029694, -0.012190070934593678, 0.026486070826649666, 0.022090423852205276, 0.09637594223022461, 0.0051352293230593204, -0.09380774945020676, -0.0015201973728835583, -0.12305431067943573, -0.01875145174562931, 0.08543036878108978, 0.02736811712384224, -0.010391109623014927, -0.0965493768453598, -0.05764399468898773, -0.14217211306095123, -0.021617302671074867, -0.02233683317899704, 0.03932022303342819, -0.053937070071697235, -0.06552916020154953, -0.04572846740484238, -0.06649041920900345, -0.05164162442088127, -0.01829097419977188, 0.11621816456317902, 0.04454735293984413, 0.007299338933080435, -0.03803687542676926, 0.11938399821519852, 0.06867919862270355, -0.12019672244787216, 0.008290603756904602, 0.008276106789708138, -0.09692328423261642, -0.054362550377845764, -0.0287662111222744, -0.017340077087283134, -0.004726592451334, 0.13373146951198578, -0.07351569086313248, 0.08250799030065536, 0.02285676822066307, -0.014526378363370895, -0.015054905787110329, 0.1435823291540146, -0.055505044758319855, -0.03784953057765961, -0.013444892130792141, 0.10780476033687592, 0.0022571219597011805, -0.013611371628940105, -0.0705476775765419, -0.0002769082202576101, 0.07124023884534836, 0.061937883496284485, -0.04537728428840637, 0.046841610223054886, -0.025883901864290237, -0.01996465027332306, 0.04160649701952934, -0.13325625658035278, 0.04336932301521301, 0.01493685133755207, -0.10206696391105652, -0.008800415322184563, 0.0151326609775424, -0.0006615486927330494, -0.04592679440975189, 0.12611831724643707, -0.07541034370660782, 0.006383739411830902, -0.08290856331586838, -0.07583671063184738, 0.011028324253857136, -0.09565798193216324, -0.03380043804645538, -0.035970695316791534, -0.1814715415239334, -0.056485366076231, 0.03608528524637222, -0.06355295330286026, -0.03852834925055504, -0.05796629190444946, -0.07574927061796188, 0.023109517991542816, -0.009658915922045708, 0.18079794943332672, -0.06615297496318817, 0.06554587185382843, -0.023169299587607384, 0.029330547899007797, 0.05839991569519043, 0.04342423379421234, -0.07115470618009567, -0.00683465925976634, -0.10657045990228653, 0.09168870747089386, -0.0807417705655098, 0.008915490470826626, -0.10842964053153992, -0.09720306098461151, 0.015182394534349442, -0.013308937661349773, 0.05966798961162567, 0.13064581155776978, -0.19953757524490356, -0.03260594978928566, 0.11414910852909088, -0.048916105180978775, -0.013492085039615631, 0.052882589399814606, -0.053741853684186935, -0.0032554687932133675, 0.05834392085671425, 0.12847530841827393, 0.06808753311634064, -0.12148094177246094, -0.007564354222267866, 0.005433284677565098, 0.03197150304913521, 0.01727290078997612, 0.01850065030157566, 0.004463187884539366, 0.06280963867902756, 0.0062888734973967075, -0.035473864525556564, 0.006256452761590481, -0.09868604689836502, -0.06884676963090897, -0.03676890209317207, -0.07424502074718475, 0.01963634602725506, 0.030668530613183975, 0.033650755882263184, -0.056541312485933304, -0.12002987414598465, 0.0875701978802681, 0.12690462172031403, -0.0589790977537632, 0.01708793081343174, -0.07594895362854004, -0.030256083235144615, 0.01003950834274292, -0.02159825898706913, -0.19784963130950928, -0.11515138298273087, 0.020878806710243225, -0.061513762921094894, 0.039850592613220215, -0.007702346425503492, 0.06367562711238861, 0.04795682057738304, -0.0397251658141613, -0.018565349280834198, -0.06479842960834503, 0.004307419527322054, -0.08937519043684006, -0.2147688865661621, -0.03397770971059799, -0.01595715619623661, 0.17363302409648895, -0.22338023781776428, 0.012245752848684788, -0.016201378777623177, 0.1512659788131714, 0.01917145401239395, -0.05881495773792267, 0.012037167325615883, 0.046344537287950516, 0.0012322074035182595, -0.09737128019332886, 0.03293362632393837, -0.005857145879417658, -0.0821346864104271, -0.02618013136088848, -0.13254451751708984, -0.027447862550616264, 0.08282434195280075, 0.07353559136390686, -0.1051352471113205, 0.01865614764392376, -0.07223834842443466, -0.04389297589659691, -0.08068595081567764, 0.036343764513731, 0.20701923966407776, 0.031879059970378876, 0.13200411200523376, -0.04803168773651123, -0.0838717669248581, -0.006255968939512968, 0.0235882755368948, 0.011110538616776466, 0.10152864456176758, 0.08850181847810745, -0.050486985594034195, 0.07380072772502899, 0.027408866211771965, -0.03830206021666527, 0.13323858380317688, -0.04294264689087868, -0.08359910547733307, 0.0011131669161841273, -0.02105332911014557, -0.026078658178448677, 0.10032155364751816, -0.048475150018930435, 0.00913478434085846, 0.030815403908491135, 0.02532307617366314, 0.03824152424931526, -0.18405303359031677, 0.003539521247148514, 0.017611611634492874, -0.056465186178684235, -0.03573494032025337, -0.022902602329850197, 0.053152862936258316, 0.1007043644785881, 0.008964749984443188, -0.03454092517495155, 0.01109483651816845, -0.005696377716958523, -0.06575586646795273, 0.1765051931142807, -0.1184539794921875, -0.1711464673280716, -0.07619491964578629, 0.03340128809213638, -0.030210023745894432, -0.047904957085847855, -0.002604816574603319, -0.11566207557916641, -0.06978494673967361, -0.10806961357593536, -0.020741498097777367, -0.002376876538619399, -0.004340469371527433, 0.05473650246858597, 0.014063458889722824, 0.056004323065280914, -0.13713373243808746, 0.017163751646876335, -0.0484476275742054, -0.08055711537599564, 0.017950447276234627, 0.08064013719558716, 0.06204807013273239, 0.15341028571128845, -0.018314067274332047, 0.027962055057287216, -0.017855199053883553, 0.1928144246339798, -0.0937369167804718, 0.013846277259290218, 0.0953863337635994, 0.015436636283993721, 0.040733274072408676, 0.12205792218446732, 0.03800421580672264, -0.08233323693275452, 0.027875611558556557, 0.09206683933734894, -0.018134960904717445, -0.25236132740974426, -0.06507786363363266, -0.017953407019376755, -0.08092175424098969, 0.08552274852991104, 0.051968421787023544, -0.008191987872123718, 0.00988634955137968, -0.007681744638830423, -0.006398170255124569, -0.0020688248332589865, 0.053261030465364456, 0.07915417850017548, 0.053145986050367355, 0.09680019319057465, -0.02765578031539917, -0.02608293481171131, 0.06385765224695206, -0.0058751655742526054, 0.2302200198173523, -0.0622570738196373, 0.07915017753839493, 0.021233487874269485, 0.09855077415704727, -0.01474290993064642, 0.02815241366624832, 0.026713302358984947, -0.011529619805514812, 0.011170346289873123, -0.04727822542190552, -0.0030879967380315065, 0.00007898466719780117, -0.012404490262269974, 0.0006275509367696941, -0.07040169090032578, 0.04532312601804733, 0.018576737493276596, 0.2655426561832428, 0.04294809326529503, -0.28906992077827454, -0.0548335500061512, -0.020548559725284576, -0.03183148428797722, -0.05417942255735397, -0.0073190126568078995, 0.0939456969499588, -0.11454855650663376, 0.06784682720899582, -0.048215627670288086, 0.08288221806287766, -0.06720715016126633, -0.00167922955006361, 0.0713920146226883, 0.14845429360866547, -0.01421202253550291, 0.06306415051221848, -0.21188032627105713, 0.21485337615013123, 0.008269239217042923, 0.12291587889194489, -0.06388825923204422, 0.026649046689271927, 0.013126501813530922, 0.04810890555381775, 0.06746691465377808, -0.004554126411676407, -0.027870317921042442, -0.15297003090381622, -0.11436145752668381, 0.028315719217061996, 0.12181136757135391, -0.002941531827673316, 0.08099119365215302, -0.027584990486502647, -0.01319347508251667, 0.043625522404909134, -0.1539730429649353, -0.16482451558113098, -0.1265249252319336, 0.033482469618320465, 0.026882203295826912, -0.03687494620680809, -0.0550273060798645, -0.11396589875221252, -0.027285899966955185, 0.18703743815422058, 0.015225808136165142, -0.058458756655454636, -0.1440007984638214, 0.07419398427009583, 0.1546095609664917, -0.039192695170640945, 0.015206653624773026, 0.030686160549521446, 0.11540747433900833, 0.052633777260780334, -0.08395690470933914, 0.04186949133872986, -0.06347525864839554, -0.18733300268650055, -0.054916467517614365, 0.1285191774368286, 0.08030910789966583, 0.04926000535488129, -0.020156873390078545, 0.027713898569345474, 0.0029754105489701033, -0.08147865533828735, 0.0074694170616567135, 0.08214516937732697, 0.08062195032835007, 0.06382518261671066, -0.0692073181271553, 0.004225732758641243, -0.03855549916625023, -0.02478344738483429, 0.10780531913042068, 0.19467920064926147, -0.08075414597988129, 0.10282517969608307, 0.0650174543261528, -0.07311355322599411, -0.1773914098739624, 0.061958566308021545, 0.12700435519218445, 0.012273470871150494, 0.027378274127840996, -0.207509845495224, 0.12475620955228806, 0.12195561081171036, -0.010575405322015285, 0.05376998335123062, -0.3446544110774994, -0.12947848439216614, 0.052897702902555466, 0.08229143917560577, 0.012945631518959999, -0.10217765718698502, -0.020623251795768738, -0.01997881755232811, -0.13449694216251373, 0.13919411599636078, -0.060084640979766846, 0.11082945019006729, -0.003810934955254197, 0.10119462013244629, 0.025122594088315964, -0.04419126734137535, 0.11206207424402237, 0.05546554923057556, 0.06988897919654846, -0.04659532755613327, 0.022150153294205666, 0.04132065922021866, -0.06326451897621155, 0.0345335379242897, -0.04165291786193848, 0.07069146633148193, -0.11303572356700897, -0.01322726160287857, -0.08359294384717941, 0.04851493239402771, -0.04564986377954483, -0.048120271414518356, -0.04231365770101547, 0.05488806962966919, 0.07640735805034637, -0.044171564280986786, 0.0635756328701973, 0.0071902647614479065, 0.0777927115559578, 0.06393005698919296, 0.09468480944633484, -0.05484994128346443, -0.07578792423009872, -0.002670957474038005, -0.008541014045476913, 0.03402605652809143, -0.11125286668539047, 0.028931286185979843, 0.14016801118850708, 0.044788114726543427, 0.12158418446779251, 0.038728099316358566, -0.03494645282626152, -0.019384684041142464, 0.03825937584042549, -0.12116114050149918, -0.09713736176490784, 0.02631295658648014, -0.07252690941095352, -0.10235565900802612, 0.018221814185380936, 0.1043429896235466, -0.03611157834529877, -0.0157100111246109, 0.0019517901819199324, 0.029671568423509598, 0.009143278934061527, 0.21174520254135132, 0.02422681637108326, 0.06405584514141083, -0.11271461099386215, 0.12897472083568573, 0.06323576718568802, -0.0851610004901886, 0.04824942722916603, 0.11608171463012695, -0.10150398313999176, -0.010186953470110893, 0.09171872586011887, 0.13689129054546356, -0.029862409457564354, -0.029733605682849884, -0.09097398817539215, -0.09422530978918076, 0.07550152391195297, 0.11813968420028687, 0.028380976989865303, 0.0026851375587284565, -0.05112950876355171, 0.019538890570402145, -0.16038207709789276, 0.06883089989423752, 0.08057249337434769, 0.05634443461894989, -0.09929938614368439, 0.14562390744686127, 0.0237188171595335, 0.029343850910663605, -0.019218558445572853, 0.016097817569971085, -0.06933943927288055, -0.011183125898241997, -0.09151063114404678, -0.017428632825613022, -0.012510299682617188, -0.004586932715028524, -0.01779460348188877, -0.04463621973991394, -0.04168948531150818, 0.04354511573910713, -0.0749322697520256, -0.05853657424449921, -0.0008325678063556552, 0.035591866821050644, -0.13716693222522736, 0.0006596326711587608, -0.0001505621476098895, -0.09235770255327225, 0.07397840917110443, 0.04680176079273224, 0.0030450185295194387, 0.03430042043328285, -0.15501806139945984, -0.04331996664404869, 0.03533318638801575, 0.0200723297894001, 0.0772152841091156, -0.076016366481781, -0.003873000852763653, -0.01788700930774212, 0.042194049805402756, 0.012601378373801708, 0.04857134819030762, -0.11037233471870422, -0.0014569731429219246, -0.0710817277431488, -0.03219319507479668, -0.06811735779047012, 0.046999741345644, 0.10560683161020279, 0.043323591351509094, 0.1762334555387497, -0.07791531085968018, 0.03630600869655609, -0.18425968289375305, -0.031272392719984055, -0.00414837384596467, -0.03685693442821503, -0.06560572236776352, -0.014584340155124664, 0.10985609143972397, -0.05799519270658493, 0.12699905037879944, 0.006683915853500366, 0.07040740549564362, 0.03975146263837814, -0.03478699550032616, -0.0628570020198822, 0.018657702952623367, 0.14926154911518097, 0.06837601959705353, -0.019871901720762253, 0.1116616353392601, -0.00694781681522727, 0.055251166224479675, 0.05072343721985817, 0.22586731612682343, 0.1308225691318512, 0.016942283138632774, 0.06982715427875519, 0.048301707953214645, -0.14455878734588623, -0.11857188493013382, 0.1233544871211052, -0.09578097611665726, 0.1285238415002823, -0.06410720199346542, 0.19993466138839722, 0.05052279680967331, -0.1467679738998413, 0.04876761883497238, -0.03820393607020378, -0.1095268577337265, -0.1093999370932579, -0.012051468715071678, -0.07920214533805847, -0.11263006925582886, 0.03434881195425987, -0.11460641771554947, 0.06256648153066635, 0.09008470922708511, 0.03244965523481369, 0.023303311318159103, 0.11122764647006989, -0.0029028896242380142, 0.00019716547103598714, 0.06079613417387009, 0.025505775585770607, -0.004513817839324474, -0.06545691937208176, -0.06990015506744385, 0.02915322035551071, 0.030470725148916245, 0.08459450304508209, -0.041605133563280106, -0.011879883706569672, 0.045039184391498566, -0.015046218410134315, -0.07169317454099655, 0.03541882336139679, 0.005546841770410538, 0.06096165254712105, 0.0462331660091877, 0.051726359874010086, 0.006126449443399906, -0.04877345636487007, 0.2851954996585846, -0.06370053440332413, -0.1104353815317154, -0.1265932321548462, 0.22831708192825317, 0.052363745868206024, -0.020400121808052063, 0.06886643916368484, -0.10472211986780167, -0.04511839151382446, 0.1531752049922943, 0.1711774617433548, -0.07059425860643387, -0.029256878420710564, -0.026764996349811554, -0.01370729599148035, -0.03589425981044769, 0.12860511243343353, 0.11202177405357361, 0.04539259523153305, -0.04753800109028816, -0.02517806552350521, -0.01649649068713188, -0.04070856049656868, -0.06995653361082077, 0.06875181198120117, 0.019518526270985603, -0.0023038999643176794, -0.035222381353378296, 0.06749159842729568, -0.013378155417740345, -0.20407193899154663, 0.03803463280200958, -0.16756878793239594, -0.1867160201072693, -0.012118091806769371, 0.09354547411203384, -0.025863531976938248, 0.044512830674648285, 0.013546470552682877, -0.015512797050178051, 0.1103498637676239, -0.02849641442298889, -0.03354806825518608, -0.09981504082679749, 0.06155984476208687, -0.1256263703107834, 0.2305879145860672, -0.007553847506642342, 0.06755364686250687, 0.09398869425058365, 0.02619466930627823, -0.12785637378692627, 0.055874962359666824, 0.04989369958639145, -0.08990154415369034, 0.0368833988904953, 0.1605667918920517, -0.04418690502643585, 0.04804104194045067, 0.03926760330796242, -0.12184792757034302, -0.0035100900568068027, -0.08383233100175858, -0.03586816415190697, -0.062155015766620636, -0.01916111633181572, -0.03280084952712059, 0.14853161573410034, 0.2174619883298874, -0.01127681601792574, 0.031233109533786774, -0.08290280401706696, -0.0011394931934773922, 0.02385796047747135, 0.09389578551054001, -0.029292868450284004, -0.2272438257932663, 0.03290386125445366, 0.01465226523578167, 0.03718313202261925, -0.17552721500396729, -0.07035580277442932, 0.00680008577182889, -0.07086638361215591, -0.05506841838359833, 0.11579443514347076, 0.04510834440588951, 0.04231797531247139, -0.02431446872651577, -0.06994646787643433, -0.01683870702981949, 0.14945518970489502, -0.18516278266906738, -0.03644043952226639 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert_gpt2_summarization_xsum This model is a fine-tuned version of [](https://huggingface.co/) on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["xsum"], "model-index": [{"name": "distilbert_gpt2_summarization_xsum", "results": []}]}
text2text-generation
Ayham/distilbert_gpt2_summarization_xsum
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:xsum", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us
# distilbert_gpt2_summarization_xsum This model is a fine-tuned version of [](URL on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# distilbert_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us \n", "# distilbert_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 59, 36, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us \n# distilbert_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.08847319334745407, 0.13946135342121124, -0.0026651325169950724, 0.06332840025424957, 0.14961284399032593, 0.04556165635585785, 0.09865406900644302, 0.11494100093841553, -0.08958621323108673, 0.08031836152076721, 0.08133233338594437, 0.07173562794923782, 0.061141010373830795, 0.1470060497522354, -0.03963296487927437, -0.24711164832115173, 0.019814638420939445, -0.02284293808043003, -0.08600224554538727, 0.09940823167562485, 0.08884028345346451, -0.11477374285459518, 0.0731453150510788, -0.004449611529707909, -0.16459187865257263, 0.008591072633862495, -0.039761465042829514, -0.04894625395536423, 0.09669999778270721, 0.0037205303087830544, 0.096199169754982, 0.02330469712615013, 0.13625235855579376, -0.2241082340478897, 0.002823382383212447, 0.09274104982614517, 0.040519196540117264, 0.08762454986572266, 0.058514244854450226, 0.0019688939210027456, 0.13073207437992096, -0.14191322028636932, 0.10251966118812561, 0.02181899920105934, -0.07775614410638809, -0.12054851651191711, -0.0934145376086235, 0.02584950253367424, 0.07948009669780731, 0.09226232767105103, 0.00584730738773942, 0.1135745570063591, -0.09807761758565903, 0.07225897908210754, 0.16737689077854156, -0.2296212762594223, -0.059305258095264435, 0.032187484204769135, 0.06194605305790901, 0.06614839285612106, -0.1062449961900711, -0.009931439533829689, 0.02118057757616043, 0.026541490107774734, 0.09203872829675674, -0.0038251203950494528, -0.09849990159273148, 0.008967310190200806, -0.12343540042638779, -0.011084876954555511, 0.10046812146902084, 0.0299794040620327, -0.026717841625213623, -0.10868091136217117, -0.057898007333278656, -0.113326296210289, -0.015088571235537529, -0.03083767369389534, 0.04467568174004555, -0.038916174322366714, -0.0722915381193161, -0.04441380500793457, -0.0648714080452919, -0.0605887845158577, -0.021246185526251793, 0.13949380815029144, 0.030931664630770683, 0.01312834769487381, -0.0341501384973526, 0.09818343073129654, 0.04054039716720581, -0.11292519420385361, -0.0029160005506128073, -0.007710671983659267, -0.11108991503715515, -0.048448070883750916, -0.05617810785770416, -0.031585708260536194, 0.012815834954380989, 0.14798672497272491, -0.06537280976772308, 0.09465935826301575, 0.02756195142865181, -0.01636767014861107, 0.003585169091820717, 0.14921143651008606, -0.051587119698524475, -0.05559786781668663, -0.019348928704857826, 0.08545117825269699, 0.014034613035619259, -0.024104690179228783, -0.0630413219332695, -0.005773062817752361, 0.07359650731086731, 0.05356060341000557, -0.0445898212492466, 0.029472004622220993, -0.04144842177629471, -0.026461752131581306, 0.0004873175057582557, -0.12416381388902664, 0.05278901383280754, 0.002671352354809642, -0.08318305015563965, 0.005112580955028534, 0.0026891762390732765, 0.010351493023335934, -0.0327720008790493, 0.128037691116333, -0.08276353031396866, -0.007590160239487886, -0.09074465185403824, -0.07862696796655655, 0.021328318864107132, -0.13597044348716736, -0.014437294565141201, -0.03511999174952507, -0.16862887144088745, -0.05068434029817581, 0.06459515541791916, -0.059275198727846146, -0.019236775115132332, -0.04959237575531006, -0.041288554668426514, 0.04111980274319649, -0.012240256182849407, 0.1627771109342575, -0.05931638926267624, 0.06932531297206879, -0.033069197088479996, 0.03862462565302849, 0.00670974887907505, 0.05048190429806709, -0.07203112542629242, 0.00871245376765728, -0.1123957633972168, 0.08640334010124207, -0.0670766830444336, 0.010671606287360191, -0.11342190206050873, -0.0861053317785263, -0.00923503004014492, -0.006271850783377886, 0.07672231644392014, 0.11999247223138809, -0.20173536241054535, -0.03312811255455017, 0.10967548191547394, -0.07079042494297028, -0.05610483139753342, 0.060091763734817505, -0.05134894698858261, 0.008335164748132229, 0.05002479627728462, 0.15914873778820038, 0.04638823866844177, -0.11721805483102798, -0.03252025321125984, 0.003796784672886133, 0.04613511636853218, 0.019771361723542213, 0.044069673866033554, -0.008772732689976692, 0.06131105124950409, 0.010427812114357948, -0.03375350683927536, 0.00530113885179162, -0.07821879535913467, -0.07415817677974701, -0.03669971972703934, -0.07200441509485245, 0.004671906121075153, 0.02498464472591877, 0.02133568376302719, -0.05893302708864212, -0.10710664093494415, 0.1108420267701149, 0.1210474967956543, -0.0756252333521843, 0.020983364433050156, -0.06731374561786652, -0.01202425081282854, -0.0008426705026067793, -0.020935405045747757, -0.20886576175689697, -0.11016187816858292, 0.027111900970339775, -0.05657588690519333, 0.03249127045273781, -0.01776527427136898, 0.06019781529903412, 0.05152919515967369, -0.03176920861005783, -0.012242327444255352, -0.06898461282253265, 0.00034248793963342905, -0.09269385039806366, -0.21739590167999268, -0.04703313112258911, -0.022486770525574684, 0.18809345364570618, -0.20444686710834503, -0.005426925141364336, -0.00046327695599757135, 0.14286044239997864, 0.02818457782268524, -0.0708802342414856, -0.002335413359105587, 0.04761875420808792, -0.010510682128369808, -0.09786994010210037, 0.03835773095488548, 0.01165727898478508, -0.09317582845687866, -0.0425686351954937, -0.15009313821792603, -0.007119121495634317, 0.09685896337032318, 0.05535281449556351, -0.08345042914152145, -0.036840248852968216, -0.06602909415960312, -0.03683701530098915, -0.07196168601512909, 0.02286873385310173, 0.18150997161865234, 0.020606862381100655, 0.11076247692108154, -0.05996052175760269, -0.07110212743282318, 0.0073294322937726974, 0.03389120101928711, -0.007839892990887165, 0.08895672112703323, 0.12709899246692657, -0.08104390650987625, 0.08282873779535294, 0.07420069724321365, -0.03176860511302948, 0.14232178032398224, -0.041387829929590225, -0.08154381066560745, -0.007058821618556976, -0.008573989383876324, -0.022283298894762993, 0.11924669146537781, -0.08815424889326096, 0.016116518527269363, 0.0285993330180645, 0.034943319857120514, 0.04342585802078247, -0.1790725737810135, 0.011810438707470894, 0.01911953277885914, -0.03676985576748848, -0.04896041005849838, -0.020735478028655052, 0.0324632003903389, 0.08870060741901398, 0.02516496554017067, 0.0066764215007424355, 0.012405935674905777, -0.008318322710692883, -0.07842620462179184, 0.18698599934577942, -0.14093129336833954, -0.1617594212293625, -0.081250861287117, 0.02716812863945961, -0.04675949737429619, -0.03244968131184578, 0.005958870518952608, -0.10493607819080353, -0.0644618347287178, -0.08932089060544968, -0.021096009761095047, -0.03532857820391655, 0.01239790115505457, 0.046694472432136536, 0.012052049860358238, 0.04576186463236809, -0.13413827121257782, 0.01188965979963541, -0.0531429760158062, -0.06157467141747475, 0.011918644420802593, 0.0801115334033966, 0.08769182115793228, 0.12933552265167236, -0.015185421332716942, 0.022902004420757294, -0.026761967688798904, 0.21326276659965515, -0.08646249771118164, 0.004401407204568386, 0.111396424472332, -0.0067789191380143166, 0.04571269452571869, 0.12035710364580154, 0.03477133437991142, -0.09608373790979385, 0.03144976124167442, 0.07387693971395493, -0.02112737111747265, -0.2434939593076706, -0.05081935226917267, -0.03833348676562309, -0.09541520476341248, 0.09854288399219513, 0.042764537036418915, -0.04610917344689369, 0.039436664432287216, -0.0037261948455125093, 0.02519424445927143, -0.015935733914375305, 0.05832091346383095, 0.08310077339410782, 0.05118013173341751, 0.10384999960660934, -0.021390898153185844, -0.014251255430281162, 0.07318703085184097, -0.002984776860103011, 0.259357213973999, -0.0359414778649807, 0.08796653151512146, 0.030101077631115913, 0.10232067853212357, -0.020913895219564438, 0.05742202699184418, 0.014870457351207733, -0.006391323171555996, -0.006764446385204792, -0.059523750096559525, -0.03678064048290253, 0.022484848275780678, -0.011395847424864769, 0.0052905394695699215, -0.09055222570896149, 0.0498289093375206, 0.03094097599387169, 0.24231773614883423, 0.026045357808470726, -0.29069244861602783, -0.05839252099394798, -0.0029296409338712692, -0.025985663756728172, -0.06826334446668625, -0.005633958149701357, 0.11645612865686417, -0.11110150068998337, 0.07509754598140717, -0.06674125045537949, 0.08459322154521942, -0.04557972028851509, -0.0019971863366663456, 0.07321041077375412, 0.16028551757335663, -0.011011440306901932, 0.057003166526556015, -0.20478495955467224, 0.21229897439479828, 0.021842071786522865, 0.12166397273540497, -0.07740132510662079, 0.04545086249709129, 0.013025105930864811, 0.014017024077475071, 0.07598842680454254, -0.0007248580805025995, -0.10673417150974274, -0.14125312864780426, -0.07825129479169846, 0.06021091714501381, 0.154796302318573, -0.011764485388994217, 0.08822160214185715, -0.044638026505708694, 0.011524539440870285, 0.04716797173023224, -0.09418012201786041, -0.15958115458488464, -0.15267765522003174, 0.028284357860684395, 0.01704881340265274, -0.027970198541879654, -0.04945361614227295, -0.09410198032855988, -0.027217036113142967, 0.16405510902404785, -0.01607939414680004, -0.04870804771780968, -0.16045089066028595, 0.07697059959173203, 0.16615864634513855, -0.05181256681680679, 0.031446423381567, 0.015187723562121391, 0.11872078478336334, 0.033445779234170914, -0.09125009179115295, 0.05789683014154434, -0.07445719093084335, -0.16311368346214294, -0.06126246973872185, 0.10510946065187454, 0.059698425233364105, 0.040564484894275665, -0.015188160352408886, 0.04104456305503845, -0.0062934523448348045, -0.09736867249011993, 0.025613291189074516, 0.09939711540937424, 0.06949024647474289, 0.06530848145484924, -0.09121225029230118, 0.03514287620782852, -0.010500801727175713, -0.022329390048980713, 0.1264263093471527, 0.1968446671962738, -0.0752883106470108, 0.10353758931159973, 0.08130767196416855, -0.08588045090436935, -0.1884491890668869, 0.08277951925992966, 0.1111360490322113, 0.023911379277706146, 0.04766964167356491, -0.22961483895778656, 0.13350632786750793, 0.12608088552951813, -0.004917416721582413, 0.05533542111515999, -0.30175912380218506, -0.12761829793453217, 0.04065016657114029, 0.1024312824010849, 0.020506631582975388, -0.11849874258041382, -0.02192125841975212, -0.04224848002195358, -0.1253897100687027, 0.14839673042297363, -0.07347609102725983, 0.11141840368509293, 0.004388649016618729, 0.07750603556632996, 0.023408882319927216, -0.04336034879088402, 0.12427441030740738, 0.027084046974778175, 0.06713234633207321, -0.036459341645240784, 0.014400246553122997, 0.005434300284832716, -0.06122547760605812, 0.02574368752539158, -0.07872039079666138, 0.06287164986133575, -0.1068522185087204, -0.015180532820522785, -0.0565839558839798, 0.05112938582897186, -0.042933300137519836, -0.06167776510119438, -0.03972171992063522, 0.04933639243245125, 0.07993762940168381, -0.03594459593296051, 0.05840533971786499, 0.0037642712704837322, 0.0799637883901596, 0.04233309254050255, 0.10310099273920059, -0.07607091218233109, -0.05945028364658356, 0.007034169044345617, -0.006557932589203119, 0.046347811818122864, -0.11449822783470154, 0.02881910651922226, 0.1410234421491623, 0.04639170318841934, 0.12979756295681, 0.057436082512140274, -0.038546882569789886, 0.001335148699581623, 0.051287028938531876, -0.11749380826950073, -0.11985968798398972, 0.014298439025878906, -0.04775587096810341, -0.12465086579322815, 0.016980599611997604, 0.11052786558866501, -0.025081602856516838, -0.0072694323025643826, -0.0134291872382164, 0.03671674057841301, 0.007469019386917353, 0.18987980484962463, 0.009307974018156528, 0.053453389555215836, -0.10857298225164413, 0.13723021745681763, 0.061045076698064804, -0.11324964463710785, 0.062443628907203674, 0.09209644049406052, -0.0889454111456871, 0.0005616546841338277, 0.0668257400393486, 0.1208706647157669, -0.03524378687143326, -0.03984663262963295, -0.08681891858577728, -0.10004615038633347, 0.05672547221183777, 0.11022264510393143, 0.028304429724812508, 0.0069373054429888725, -0.04709972068667412, 0.03483140841126442, -0.1515703946352005, 0.0649145171046257, 0.05368392542004585, 0.06315968930721283, -0.11316267400979996, 0.11799326539039612, 0.017224149778485298, 0.009404953569173813, -0.014567933045327663, 0.018389558419585228, -0.09763307124376297, -0.026263341307640076, -0.08292298763990402, -0.015639303252100945, -0.03863846883177757, -0.0021952453535050154, -0.005265187006443739, -0.0379578098654747, -0.053582631051540375, 0.03256170451641083, -0.0733889788389206, -0.06062255799770355, -0.005926834885030985, 0.038202714174985886, -0.13947689533233643, 0.007827211171388626, 0.013255405239760876, -0.10008441656827927, 0.08227375149726868, 0.054104067385196686, 0.01643470861017704, 0.03590116649866104, -0.15750135481357574, -0.04732520878314972, 0.02848675660789013, 0.0223518256098032, 0.0708879828453064, -0.08816851675510406, -0.008959867991507053, -0.007530813105404377, 0.044649943709373474, 0.010351140983402729, 0.0563882440328598, -0.11182717233896255, 0.0011659519514068961, -0.07459748536348343, -0.04930799454450607, -0.062353771179914474, 0.04753196984529495, 0.11470606923103333, 0.04102277755737305, 0.1651592254638672, -0.06517720967531204, 0.033589527010917664, -0.1901370733976364, -0.023152682930231094, -0.006801077164709568, -0.03317352011799812, -0.08835238963365555, -0.023572582751512527, 0.09451402723789215, -0.05046803876757622, 0.13033807277679443, 0.008058690465986729, 0.06887449324131012, 0.03655123710632324, -0.027808377519249916, -0.051034387201070786, 0.013485906645655632, 0.19397549331188202, 0.0785960853099823, -0.020470624789595604, 0.09323006123304367, 0.004778036382049322, 0.060599349439144135, 0.06261809915304184, 0.23615044355392456, 0.15500599145889282, 0.012676416896283627, 0.0832541361451149, 0.0540502667427063, -0.1287965476512909, -0.12472721934318542, 0.15630613267421722, -0.06347540020942688, 0.12318681925535202, -0.053278692066669464, 0.19086571037769318, 0.05482077598571777, -0.1715690791606903, 0.03539801388978958, -0.04330763965845108, -0.10267923027276993, -0.12753455340862274, -0.00210542231798172, -0.08340662717819214, -0.11788048595190048, 0.022144915536046028, -0.12946712970733643, 0.0583987757563591, 0.08374752849340439, 0.01886218599975109, 0.0329376719892025, 0.10711731016635895, -0.031161224469542503, 0.0029904821421951056, 0.05091287940740585, 0.036017097532749176, -0.0023995775263756514, -0.042571380734443665, -0.08937296271324158, 0.030816948041319847, 0.015812216326594353, 0.09445764124393463, -0.04647376388311386, -0.015158367343246937, 0.048710763454437256, -0.010217606090009212, -0.06346289813518524, 0.02125205285847187, 0.006906371563673019, 0.043022483587265015, 0.042732562869787216, 0.045654915273189545, -0.0060422830283641815, -0.04794730618596077, 0.26834267377853394, -0.0661228597164154, -0.07171311229467392, -0.1348988562822342, 0.20090104639530182, 0.03808337450027466, -0.012798802927136421, 0.07137922942638397, -0.10729220509529114, -0.04270673170685768, 0.16005422174930573, 0.15575218200683594, -0.08539853245019913, -0.030349591746926308, -0.010520615614950657, -0.009557397104799747, -0.05124915763735771, 0.14255717396736145, 0.09595967829227448, 0.07444093376398087, -0.048304565250873566, -0.01917407475411892, -0.017969315871596336, -0.04569929465651512, -0.0565001517534256, 0.056355513632297516, 0.020207254216074944, -0.009986829943954945, -0.02826545760035515, 0.07047172635793686, -0.00924069993197918, -0.18531665205955505, 0.038936082273721695, -0.18007682263851166, -0.17862822115421295, -0.02028467133641243, 0.09279482066631317, -0.028164584189653397, 0.04889499023556709, -0.0008964557200670242, -0.017978129908442497, 0.10583376884460449, -0.008815353736281395, -0.048265017569065094, -0.09443347901105881, 0.07683952152729034, -0.06698036938905716, 0.2215435802936554, -0.010974396020174026, 0.06185796856880188, 0.10153265297412872, 0.05603170394897461, -0.12808047235012054, 0.025889357551932335, 0.05911997705698013, -0.1075020432472229, 0.0373581238090992, 0.138715922832489, -0.06195363402366638, 0.07445953041315079, 0.048256777226924896, -0.10898882150650024, 0.005979916080832481, -0.06346461176872253, -0.042584169656038284, -0.05637732520699501, -0.019370516762137413, -0.06587374955415726, 0.1649647206068039, 0.20912985503673553, -0.013957557268440723, 0.01830102689564228, -0.07497616857290268, 0.03314455971121788, 0.028672315180301666, 0.1120990514755249, -0.037256162613630295, -0.23372089862823486, 0.026353025808930397, 0.02868015691637993, 0.013866403140127659, -0.1911264806985855, -0.0759928971529007, 0.029304491356015205, -0.05051634833216667, -0.06362181901931763, 0.10791807621717453, 0.057980626821517944, 0.02530308999121189, -0.04021920636296272, -0.10025161504745483, -0.024046342819929123, 0.14606468379497528, -0.14943796396255493, -0.040356505662202835 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert_roberta_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "distilbert_roberta_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/distilbert_roberta_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# distilbert_roberta_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
[ "# distilbert_roberta_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# distilbert_roberta_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ 62, 41, 6, 12, 8, 3, 118, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# distilbert_roberta_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ -0.08382948487997055, 0.1381072700023651, -0.003844823921099305, 0.07354782521724701, 0.1329762041568756, 0.040541648864746094, 0.11363682895898819, 0.1209488958120346, -0.11866962164640427, 0.057980168610811234, 0.07231302559375763, 0.06643041968345642, 0.04707719013094902, 0.15298326313495636, -0.029337236657738686, -0.22638866305351257, 0.0031102639622986317, -0.004687830340117216, -0.07596194744110107, 0.11801969259977341, 0.08519414812326431, -0.10124070942401886, 0.06188303232192993, -0.003846007864922285, -0.13452517986297607, 0.01910185068845749, -0.04239365831017494, -0.05151202157139778, 0.09674874693155289, -0.01509255450218916, 0.09376190602779388, 0.031199831515550613, 0.14426231384277344, -0.2015528380870819, 0.003910257946699858, 0.09521273523569107, 0.036143943667411804, 0.09224783629179001, 0.06384241580963135, -0.013468512333929539, 0.1362404078245163, -0.1676226556301117, 0.10289791971445084, 0.011025763116776943, -0.08000004291534424, -0.10590257495641708, -0.10176081955432892, 0.057210054248571396, 0.09351109713315964, 0.11531073600053787, 0.006340265739709139, 0.132041797041893, -0.10857980698347092, 0.08351128548383713, 0.16944722831249237, -0.20822717249393463, -0.053328290581703186, 0.06701401621103287, 0.021703630685806274, 0.05744586139917374, -0.09069675952196121, -0.01676436886191368, 0.022093961015343666, 0.024779897183179855, 0.10065718740224838, -0.00046048691729083657, -0.0999450534582138, -0.00011240693129366264, -0.1264907568693161, -0.02377605065703392, 0.08273223042488098, 0.016418928280472755, -0.009019180200994015, -0.09575392305850983, -0.055115241557359695, -0.13912788033485413, -0.02294657938182354, -0.018434302881360054, 0.04032445698976517, -0.05752842128276825, -0.06913311034440994, -0.04414352774620056, -0.0692177414894104, -0.049291521310806274, -0.027963608503341675, 0.11302082240581512, 0.04487299174070358, 0.011385512538254261, -0.05185224860906601, 0.11904596537351608, 0.0673849806189537, -0.12661252915859222, 0.010905052535235882, 0.01339578628540039, -0.09261130541563034, -0.056578606367111206, -0.025518471375107765, -0.034131817519664764, -0.010616487823426723, 0.13852110505104065, -0.07532887160778046, 0.07847325503826141, 0.024250058457255363, -0.017813457176089287, -0.010757296346127987, 0.13670197129249573, -0.048080991953611374, -0.04228176549077034, -0.013363736681640148, 0.10722597688436508, 0.006663942243903875, -0.010095376521348953, -0.06692629307508469, 0.00011858815560117364, 0.07116682827472687, 0.06014464423060417, -0.03557494282722473, 0.044914063066244125, -0.016106367111206055, -0.02002290077507496, 0.04187248647212982, -0.13641230762004852, 0.04263374209403992, 0.014629889279603958, -0.10180693119764328, -0.004988113883882761, 0.013821202330291271, -0.0014991858042776585, -0.04786786064505577, 0.1382371038198471, -0.07852473855018616, 0.011018604040145874, -0.08013186603784561, -0.07141247391700745, 0.017485054209828377, -0.10300005227327347, -0.0407869778573513, -0.032068461179733276, -0.1931891292333603, -0.05854228883981705, 0.026711437851190567, -0.05910707637667656, -0.034132178872823715, -0.058345988392829895, -0.07287688553333282, 0.02066725865006447, -0.005828586872667074, 0.18056495487689972, -0.06794673204421997, 0.06604625284671783, -0.027194133028388023, 0.02601870708167553, 0.06680695712566376, 0.04336319863796234, -0.07083483785390854, -0.005138642154633999, -0.10103654116392136, 0.08907418698072433, -0.0738845244050026, 0.006299844942986965, -0.10677918046712875, -0.09697028994560242, 0.015335818752646446, -0.013341899029910564, 0.05296564847230911, 0.13746297359466553, -0.20501360297203064, -0.03251614421606064, 0.11940552294254303, -0.05878311023116112, -0.009564507752656937, 0.053834401071071625, -0.061215318739414215, -0.0003738359082490206, 0.06884420663118362, 0.12912020087242126, 0.06455350667238235, -0.1270173192024231, -0.010696305893361568, -0.007787258829921484, 0.031132027506828308, 0.008239587768912315, 0.018332256004214287, 0.009396336041390896, 0.06638137251138687, 0.004936902783811092, -0.03984164446592331, 0.005700584966689348, -0.0950060486793518, -0.06806793808937073, -0.037636566907167435, -0.07185570895671844, 0.015994248911738396, 0.026606649160385132, 0.03732674568891525, -0.05649580806493759, -0.10789211094379425, 0.09486711025238037, 0.12435310333967209, -0.05222352221608162, 0.014766877517104149, -0.070053331553936, -0.038380756974220276, 0.011283613741397858, -0.024156907573342323, -0.19817692041397095, -0.11016736179590225, 0.021935468539595604, -0.06012484431266785, 0.04324113205075264, -0.011904573999345303, 0.06842122972011566, 0.041325971484184265, -0.0400104857981205, -0.01564480923116207, -0.06108913943171501, 0.0012270156294107437, -0.08929834514856339, -0.20562098920345306, -0.029670892283320427, -0.018492072820663452, 0.15486273169517517, -0.21140886843204498, 0.015259959734976292, -0.016465753316879272, 0.15154515206813812, 0.025110343471169472, -0.05809642747044563, 0.019562790170311928, 0.052636630833148956, 0.00896112434566021, -0.09397979080677032, 0.03003218024969101, -0.010233893990516663, -0.0711415633559227, -0.021552784368395805, -0.13717956840991974, -0.030340120196342468, 0.07722357660531998, 0.08904664218425751, -0.10592533648014069, 0.016615983098745346, -0.07387711107730865, -0.048395294696092606, -0.09242912381887436, 0.0333145447075367, 0.2024623155593872, 0.03543531149625778, 0.1271587461233139, -0.04939764365553856, -0.08390924334526062, -0.01022516842931509, 0.02456364594399929, 0.006730503868311644, 0.09947173297405243, 0.08113710582256317, -0.04545682296156883, 0.07382981479167938, 0.016110850498080254, -0.045955609530210495, 0.13694749772548676, -0.04680866003036499, -0.08755646646022797, -0.005822590086609125, -0.028484756126999855, -0.02427937276661396, 0.09865568578243256, -0.03368714824318886, 0.005745639093220234, 0.03218456730246544, 0.02723049372434616, 0.0350748747587204, -0.18246684968471527, -0.0023838630877435207, 0.012682539410889149, -0.05818092077970505, -0.031376659870147705, -0.013991118408739567, 0.05422293394804001, 0.10257956385612488, 0.0036899240221828222, -0.032075315713882446, 0.011705439537763596, -0.006589507218450308, -0.07078810781240463, 0.17554086446762085, -0.11497867107391357, -0.16500410437583923, -0.08208361268043518, 0.02725195325911045, -0.033969566226005554, -0.05174477398395538, 0.002915613818913698, -0.1104666218161583, -0.06815911829471588, -0.11249957233667374, -0.02120821177959442, 0.006820546463131905, -0.0004169073945377022, 0.036773931235075, 0.015411920845508575, 0.04752320423722267, -0.13669130206108093, 0.013772977516055107, -0.05059535801410675, -0.07764452695846558, 0.02060301974415779, 0.07653889805078506, 0.059695567935705185, 0.14682409167289734, -0.01334807276725769, 0.0294630266726017, -0.016815446317195892, 0.1969611793756485, -0.09135434776544571, 0.018304407596588135, 0.10320402681827545, 0.017539147287607193, 0.0448235422372818, 0.1222788617014885, 0.040164802223443985, -0.07291638106107712, 0.025731228291988373, 0.09303677082061768, -0.020973604172468185, -0.26246708631515503, -0.06336185336112976, -0.016367202624678612, -0.0805714949965477, 0.08795614540576935, 0.0583292581140995, -0.0013953166780993342, 0.009384188801050186, -0.007904404774308205, -0.01388554647564888, -0.006484969984740019, 0.054739080369472504, 0.08382134139537811, 0.05710447207093239, 0.0897556021809578, -0.03199946880340576, -0.03251440450549126, 0.06452658027410507, -0.004522278904914856, 0.22312061488628387, -0.059634365141391754, 0.07331155240535736, 0.018757684156298637, 0.0888466015458107, -0.01769210584461689, 0.032972369343042374, 0.025555571541190147, -0.009549706242978573, 0.01477325800806284, -0.05222950130701065, 0.0019311547512188554, 0.00248295278288424, -0.01544055063277483, -0.0031741929706186056, -0.05629825219511986, 0.037899937480688095, 0.019031982868909836, 0.2789958715438843, 0.03751443699002266, -0.2857675552368164, -0.04797888547182083, -0.021633196622133255, -0.03151150047779083, -0.04622865468263626, -0.013671775348484516, 0.09302424639463425, -0.11445218324661255, 0.07032901793718338, -0.04712029546499252, 0.0835442841053009, -0.07654643803834915, -0.0031436264980584383, 0.0634026974439621, 0.1379285603761673, -0.019462142139673233, 0.05845387279987335, -0.21827127039432526, 0.2182217538356781, 0.011647404171526432, 0.1281414031982422, -0.06885918229818344, 0.02066885121166706, 0.017214106395840645, 0.04441995173692703, 0.06662135571241379, -0.005260947160422802, -0.027641495689749718, -0.14351963996887207, -0.1295853853225708, 0.023283662274479866, 0.11115501075983047, 0.00005457939187181182, 0.07787162810564041, -0.01646503247320652, -0.016061685979366302, 0.036657609045505524, -0.16076086461544037, -0.1593327820301056, -0.11634289473295212, 0.03783606365323067, 0.03201929107308388, -0.04457033798098564, -0.05201509967446327, -0.12065041065216064, -0.02580789104104042, 0.16949822008609772, 0.0261098463088274, -0.06319243460893631, -0.14308512210845947, 0.056723546236753464, 0.15487462282180786, -0.04078802838921547, 0.011061804369091988, 0.03658555448055267, 0.11271946132183075, 0.04302830621600151, -0.0853927880525589, 0.029640939086675644, -0.05913426727056503, -0.1845688670873642, -0.05519174784421921, 0.13132403790950775, 0.0895361676812172, 0.046985916793346405, -0.020669855177402496, 0.03269963711500168, -0.0011191396042704582, -0.07977035641670227, 0.002507605589926243, 0.07760757952928543, 0.08675703406333923, 0.06042508780956268, -0.06611030548810959, -0.006403931882232428, -0.04545331746339798, -0.020695483312010765, 0.09757128357887268, 0.1858041137456894, -0.07674713432788849, 0.10457714647054672, 0.059966400265693665, -0.0671355351805687, -0.16995906829833984, 0.056470632553100586, 0.1272597461938858, 0.009878062643110752, 0.037267353385686874, -0.20179197192192078, 0.11845362186431885, 0.1236359179019928, -0.011029248125851154, 0.05169733986258507, -0.3548945486545563, -0.12805773317813873, 0.050500206649303436, 0.08241946995258331, 0.015923133119940758, -0.10560119897127151, -0.02158854342997074, -0.006289216224104166, -0.1322721242904663, 0.14642493426799774, -0.067635677754879, 0.1088816225528717, -0.011517932638525963, 0.11271165311336517, 0.024528346955776215, -0.04681600630283356, 0.11167352646589279, 0.07421412318944931, 0.07258966565132141, -0.048832837492227554, 0.009062603116035461, 0.04443185403943062, -0.07010043412446976, 0.045054417103528976, -0.04355255514383316, 0.06891772150993347, -0.1288648396730423, -0.01277865655720234, -0.08321883529424667, 0.044051725417375565, -0.04702033847570419, -0.044877395033836365, -0.03908364847302437, 0.05509797856211662, 0.07239460200071335, -0.046156417578458786, 0.06994778662919998, 0.009356733411550522, 0.08035904914140701, 0.07075279206037521, 0.08604447543621063, -0.04128580167889595, -0.09551376849412918, -0.00854312814772129, -0.008300086483359337, 0.030782494693994522, -0.1204148605465889, 0.02017717808485031, 0.13660934567451477, 0.049575164914131165, 0.12452836334705353, 0.03143071010708809, -0.03871340677142143, -0.0212317556142807, 0.03711515665054321, -0.12950865924358368, -0.08654141426086426, 0.031753286719322205, -0.07248416543006897, -0.10832751542329788, 0.013421413488686085, 0.1002577617764473, -0.03595961630344391, -0.018581818789243698, 0.006222765892744064, 0.03040255978703499, 0.010912338271737099, 0.21901649236679077, 0.025487195700407028, 0.06095632538199425, -0.1135161817073822, 0.12562640011310577, 0.06660366803407669, -0.08611225336790085, 0.04399408400058746, 0.12849022448062897, -0.09936222434043884, -0.012843374162912369, 0.1007545217871666, 0.1471959799528122, -0.03199506178498268, -0.030909178778529167, -0.09175397455692291, -0.08842647075653076, 0.07853686809539795, 0.12780702114105225, 0.032070260494947433, 0.00783368106931448, -0.055333543568849564, 0.010961725376546383, -0.1458742618560791, 0.06660142540931702, 0.09013654291629791, 0.05771331116557121, -0.08180699497461319, 0.1426210254430771, 0.022933531552553177, 0.03480379655957222, -0.01749950274825096, 0.013244441710412502, -0.06895172595977783, -0.012165564112365246, -0.07946930825710297, -0.009913591668009758, -0.016046015545725822, -0.003637885907664895, -0.020153405144810677, -0.05031167343258858, -0.04219912365078926, 0.040973469614982605, -0.07208845019340515, -0.05425504595041275, -0.004276209510862827, 0.039275873452425, -0.13768929243087769, -0.0005420165252871811, 0.0011518491664901376, -0.09167376160621643, 0.07124024629592896, 0.04982483386993408, -0.004399663768708706, 0.029484868049621582, -0.14592425525188446, -0.034910157322883606, 0.030051354318857193, 0.01579137146472931, 0.07451774179935455, -0.07404068857431412, -0.0037014330737292767, -0.01906396448612213, 0.036723870784044266, 0.018324999138712883, 0.04622757434844971, -0.11352857947349548, 0.016500083729624748, -0.06839343905448914, -0.03375084325671196, -0.06795094162225723, 0.03919446840882301, 0.11253909021615982, 0.039659060537815094, 0.1768811196088791, -0.07902771234512329, 0.04002773389220238, -0.18259397149085999, -0.030611369758844376, -0.003077808767557144, -0.042018067091703415, -0.06613817065954208, -0.0140611557289958, 0.10944515466690063, -0.05303483083844185, 0.12360407412052155, 0.014936132356524467, 0.06074183061718941, 0.04025765135884285, -0.02238425426185131, -0.06838394701480865, 0.011876753531396389, 0.1418992429971695, 0.059423305094242096, -0.021152159199118614, 0.11244579404592514, -0.011974030174314976, 0.04597177356481552, 0.047500886023044586, 0.2282523512840271, 0.1272398829460144, 0.02410542406141758, 0.0685776025056839, 0.05348866805434227, -0.13219782710075378, -0.11286014318466187, 0.10491451621055603, -0.08741197735071182, 0.11862812936306, -0.06342373788356781, 0.19899696111679077, 0.055467721074819565, -0.14901410043239594, 0.05488047003746033, -0.022796165198087692, -0.10381725430488586, -0.1020151749253273, -0.015498784370720387, -0.07640840113162994, -0.1002546101808548, 0.03385182097554207, -0.11190205067396164, 0.06303147226572037, 0.09340982884168625, 0.03425261750817299, 0.020173991098999977, 0.11477024108171463, -0.0037251850590109825, -0.0002976806426886469, 0.0623934343457222, 0.024842362850904465, -0.00030699282069690526, -0.06588587909936905, -0.06454475969076157, 0.020556466653943062, 0.027745423838496208, 0.0829247534275055, -0.03694917634129524, 0.0035941817332059145, 0.040483616292476654, -0.018841225653886795, -0.07422842085361481, 0.038830313831567764, 0.004027046263217926, 0.06593981385231018, 0.05816638097167015, 0.058477722108364105, 0.007624457590281963, -0.045879971235990524, 0.2770872712135315, -0.057286329567432404, -0.1107771024107933, -0.1234666183590889, 0.22081497311592102, 0.05075136199593544, -0.026856958866119385, 0.06410819292068481, -0.1072588711977005, -0.04371983930468559, 0.15476052463054657, 0.17280550301074982, -0.053926292806863785, -0.025463512167334557, -0.03376541659235954, -0.010386702604591846, -0.019357187673449516, 0.1156800389289856, 0.09986122697591782, 0.053613100200891495, -0.04837355762720108, -0.020301762968301773, -0.01913808286190033, -0.03885725885629654, -0.0758252739906311, 0.06237323582172394, 0.02008516900241375, -0.007918949238955975, -0.03265795111656189, 0.07793932408094406, -0.018468376249074936, -0.20076794922351837, 0.04214105010032654, -0.17152954638004303, -0.18724513053894043, -0.003676322288811207, 0.09827665984630585, -0.025364242494106293, 0.04643634334206581, 0.013970104977488518, -0.015973534435033798, 0.097418412566185, -0.030573828145861626, -0.037460267543792725, -0.10592030733823776, 0.06879536807537079, -0.11688363552093506, 0.22321976721286774, -0.004276549443602562, 0.07601595669984818, 0.08924458175897598, 0.022499706596136093, -0.12661540508270264, 0.05994313955307007, 0.04438600689172745, -0.09151872247457504, 0.029383325949311256, 0.15946641564369202, -0.04233771190047264, 0.031481508165597916, 0.03609244152903557, -0.110383041203022, -0.006384717300534248, -0.0755777433514595, -0.02380942739546299, -0.06886376440525055, -0.013464073650538921, -0.026001667603850365, 0.1503811776638031, 0.2122102975845337, -0.00671009998768568, 0.03181888908147812, -0.07864322513341904, -0.004409638233482838, 0.02512354962527752, 0.0764067992568016, -0.030922450125217438, -0.22403500974178314, 0.03327367454767227, 0.008672760799527168, 0.038535226136446, -0.17473725974559784, -0.07411254197359085, 0.009596378542482853, -0.0720684751868248, -0.05202893912792206, 0.1077960729598999, 0.04178363457322121, 0.039796315133571625, -0.020737072452902794, -0.07336623221635818, -0.010747246444225311, 0.14611129462718964, -0.19190023839473724, -0.03745950013399124 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ernie_gpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "ernie_gpt2_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/ernie_gpt2_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# ernie_gpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
[ "# ernie_gpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# ernie_gpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ 62, 41, 6, 12, 8, 3, 118, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# ernie_gpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ -0.08359525352716446, 0.1347220093011856, -0.0038325923960655928, 0.075215183198452, 0.13660329580307007, 0.041458409279584885, 0.11215867847204208, 0.12396219372749329, -0.11001863330602646, 0.059094373136758804, 0.07114836573600769, 0.06988802552223206, 0.04827146232128143, 0.15441285073757172, -0.024536145851016045, -0.22868789732456207, 0.0026855897158384323, -0.003930998966097832, -0.07158468663692474, 0.11625594645738602, 0.0836915522813797, -0.09828859567642212, 0.06344572454690933, -0.003482835367321968, -0.13012148439884186, 0.01579221338033676, -0.0423898920416832, -0.0504976250231266, 0.0977204293012619, -0.014406155794858932, 0.08479197323322296, 0.028857167810201645, 0.14496289193630219, -0.206111878156662, 0.0028929635882377625, 0.09670618176460266, 0.03605711832642555, 0.0908779501914978, 0.061941228806972504, -0.010417057201266289, 0.12910059094429016, -0.1724981963634491, 0.10179933160543442, 0.01234367024153471, -0.08079993724822998, -0.11318064481019974, -0.09787581861019135, 0.05067254975438118, 0.09625472128391266, 0.1161893829703331, 0.00326903210952878, 0.12818802893161774, -0.10814125090837479, 0.07989411056041718, 0.17357513308525085, -0.20716969668865204, -0.047254737466573715, 0.06941518932580948, 0.02458692528307438, 0.05701176077127457, -0.08944375813007355, -0.017000332474708557, 0.022964736446738243, 0.02345097064971924, 0.10100623220205307, 0.0013740223366767168, -0.09068295359611511, 0.0003462469903752208, -0.12860126793384552, -0.025590762495994568, 0.07424578070640564, 0.018345417454838753, -0.010123122483491898, -0.10317595303058624, -0.05483042448759079, -0.14025218784809113, -0.022061971947550774, -0.018362266942858696, 0.04079509899020195, -0.057073548436164856, -0.06555692106485367, -0.04605104774236679, -0.06681611388921738, -0.052540093660354614, -0.022745242342352867, 0.11323458701372147, 0.04583952575922012, 0.012632624246180058, -0.05051669478416443, 0.12357980012893677, 0.05869632214307785, -0.12656116485595703, 0.009346464648842812, 0.017026297748088837, -0.09157893806695938, -0.055027034133672714, -0.021265696734189987, -0.024651145562529564, -0.011445986106991768, 0.13812576234340668, -0.07363639026880264, 0.08070109784603119, 0.030758630484342575, -0.017277538776397705, -0.009018843062222004, 0.13896909356117249, -0.04704302176833153, -0.03730634227395058, -0.01227970328181982, 0.10833064466714859, 0.007027659565210342, -0.014101693406701088, -0.07002774626016617, -0.002315943595021963, 0.06980860233306885, 0.06220366060733795, -0.03933270275592804, 0.046022459864616394, -0.020138312131166458, -0.019510330632328987, 0.04458732157945633, -0.13521282374858856, 0.0462038591504097, 0.015691708773374557, -0.10154788196086884, -0.0076014515943825245, 0.009242759086191654, -0.0009172970312647521, -0.0502784438431263, 0.13578039407730103, -0.07936739921569824, 0.011768937110900879, -0.07868024706840515, -0.07040318846702576, 0.018181396648287773, -0.10610257089138031, -0.03732607886195183, -0.03279101103544235, -0.19413349032402039, -0.05742086470127106, 0.024201957508921623, -0.059517767280340195, -0.03724746033549309, -0.06278581917285919, -0.07138977199792862, 0.019979413598775864, -0.006638981867581606, 0.18270696699619293, -0.07144604623317719, 0.06531162559986115, -0.02622167579829693, 0.02706935815513134, 0.07073833048343658, 0.04145151004195213, -0.07016527652740479, -0.0036585847847163677, -0.10301987081766129, 0.09172007441520691, -0.07369839400053024, 0.00007591504254378378, -0.10692496597766876, -0.09709488600492477, 0.024091672152280807, -0.015104521997272968, 0.05272996798157692, 0.14065055549144745, -0.1989918053150177, -0.032559849321842194, 0.11998319625854492, -0.05446143075823784, -0.0047832694835960865, 0.05962756648659706, -0.05962555482983589, -0.001220491947606206, 0.06978360563516617, 0.13308271765708923, 0.06438350677490234, -0.12979461252689362, -0.012057912535965443, -0.005037783645093441, 0.0301048681139946, 0.012009281665086746, 0.02114091068506241, 0.007428111508488655, 0.06296113133430481, 0.004168398678302765, -0.039960335940122604, 0.006890967022627592, -0.09602568298578262, -0.06634984910488129, -0.038401585072278976, -0.07052472978830338, 0.022468408569693565, 0.02293248474597931, 0.03340655192732811, -0.05585145205259323, -0.11181880533695221, 0.08916635811328888, 0.1228073462843895, -0.05443955585360527, 0.014398038387298584, -0.073207788169384, -0.03387860953807831, 0.01000087521970272, -0.02268097549676895, -0.1910155713558197, -0.11413204669952393, 0.02568683587014675, -0.06890985369682312, 0.040983691811561584, -0.014286127872765064, 0.06887796521186829, 0.045362409204244614, -0.0390080027282238, -0.016355428844690323, -0.06571083515882492, -0.0014596364926546812, -0.09327930212020874, -0.20043033361434937, -0.03168503940105438, -0.018192924559116364, 0.15247854590415955, -0.21389149129390717, 0.016740767285227776, -0.01854286715388298, 0.14421901106834412, 0.02424577809870243, -0.06010152027010918, 0.018424563109874725, 0.04938070848584175, 0.010142149403691292, -0.09606070816516876, 0.031182510778307915, -0.010542322881519794, -0.07366447150707245, -0.018940752372145653, -0.14251264929771423, -0.03157856687903404, 0.07321186363697052, 0.09184007346630096, -0.10338599234819412, 0.014415983110666275, -0.0730929747223854, -0.04897972196340561, -0.10093914717435837, 0.031491469591856, 0.21452811360359192, 0.03556520491838455, 0.12679368257522583, -0.04912274703383446, -0.08545155078172684, -0.0069910879246890545, 0.02442944422364235, 0.009087542071938515, 0.1001533716917038, 0.0839788094162941, -0.03244280442595482, 0.07309655845165253, 0.01204557903110981, -0.043290287256240845, 0.13816595077514648, -0.044444646686315536, -0.08881007879972458, -0.006010445300489664, -0.03004510886967182, -0.026369618251919746, 0.09806445240974426, -0.044124800711870193, 0.007038460113108158, 0.03204813599586487, 0.03145408630371094, 0.033570796251297, -0.18642453849315643, -0.004065518267452717, 0.015201537869870663, -0.06285206973552704, -0.03722869232296944, -0.013880278915166855, 0.05186945199966431, 0.1011398509144783, 0.0038311847019940615, -0.029294254258275032, 0.012778733856976032, -0.004995287396013737, -0.07309135049581528, 0.176283061504364, -0.1100398525595665, -0.16260913014411926, -0.08351805806159973, 0.027344919741153717, -0.029401645064353943, -0.05091357231140137, 0.005259518511593342, -0.1123952642083168, -0.06681779772043228, -0.1146666407585144, -0.016347507014870644, 0.007744867820292711, -0.0018624605145305395, 0.0348375029861927, 0.011846617795526981, 0.04494381323456764, -0.137091264128685, 0.011485922150313854, -0.04627068340778351, -0.08390025794506073, 0.019331488758325577, 0.07999490946531296, 0.055827315896749496, 0.15077345073223114, -0.012914685532450676, 0.028450535610318184, -0.017955297604203224, 0.20452071726322174, -0.09402219206094742, 0.017014330253005028, 0.10563671588897705, 0.019160868600010872, 0.04664909094572067, 0.11849047988653183, 0.03923889622092247, -0.0750652551651001, 0.025839287787675858, 0.09441623091697693, -0.023184707388281822, -0.2625127136707306, -0.060282908380031586, -0.012083521112799644, -0.07424286007881165, 0.08828412741422653, 0.057974182069301605, -0.006512546446174383, 0.012254372239112854, -0.006915138103067875, -0.009807638823986053, -0.008417584002017975, 0.05506573244929314, 0.0778127908706665, 0.056248560547828674, 0.08505139499902725, -0.028181178495287895, -0.035258226096630096, 0.0637834444642067, 0.002047258894890547, 0.22066955268383026, -0.05921626463532448, 0.08114010840654373, 0.014635946601629257, 0.0892631933093071, -0.017856469377875328, 0.03127405792474747, 0.02590232342481613, -0.010151877999305725, 0.013474708423018456, -0.05052462965250015, 0.004881754517555237, 0.0023672825191169977, -0.0105692557990551, -0.00682272482663393, -0.06176096200942993, 0.03670000657439232, 0.018431857228279114, 0.27712303400039673, 0.040800031274557114, -0.2861177325248718, -0.04603312909603119, -0.02069796621799469, -0.032091423869132996, -0.049242496490478516, -0.01318609993904829, 0.09738556295633316, -0.1187136098742485, 0.06851735711097717, -0.046820878982543945, 0.08404866605997086, -0.08296958357095718, -0.0026025488041341305, 0.06280961632728577, 0.1362721025943756, -0.021586216986179352, 0.06261671334505081, -0.21962517499923706, 0.21170583367347717, 0.011786685325205326, 0.12960755825042725, -0.0726027712225914, 0.01845437102019787, 0.0155520373955369, 0.03875327855348587, 0.07038101553916931, -0.004632036201655865, -0.026606179773807526, -0.15419301390647888, -0.1352570503950119, 0.026948919519782066, 0.10919615626335144, -0.0038003860972821712, 0.07603473216295242, -0.01628531888127327, -0.01605173572897911, 0.03408583998680115, -0.1577901840209961, -0.1581735461950302, -0.11272072046995163, 0.03625163063406944, 0.032386232167482376, -0.03748674318194389, -0.04932600259780884, -0.12045988440513611, -0.025863079354166985, 0.16622623801231384, 0.028001785278320312, -0.06523190438747406, -0.1403508484363556, 0.050511185079813004, 0.1524384319782257, -0.041104286909103394, 0.012029901146888733, 0.04206301271915436, 0.11480288952589035, 0.045833393931388855, -0.08423934131860733, 0.03242606297135353, -0.05720027908682823, -0.18996070325374603, -0.0554131343960762, 0.13167549669742584, 0.09344052523374557, 0.047788359224796295, -0.02059103175997734, 0.029748158529400826, -0.000036242032365407795, -0.08347338438034058, 0.009504240937530994, 0.07927466183900833, 0.0875382050871849, 0.06296250969171524, -0.06195983663201332, 0.0018845058511942625, -0.04904024675488472, -0.02338951826095581, 0.10000262409448624, 0.19219088554382324, -0.07615011930465698, 0.09956927597522736, 0.06314093619585037, -0.0673639327287674, -0.1669052541255951, 0.05495969206094742, 0.12898491322994232, 0.010899227112531662, 0.03420991823077202, -0.2042863965034485, 0.11625804752111435, 0.12951527535915375, -0.011415364220738411, 0.06726711243391037, -0.3629034757614136, -0.13153080642223358, 0.04858126863837242, 0.0828956738114357, 0.01892992854118347, -0.10485906153917313, -0.020289763808250427, -0.004677068442106247, -0.13088355958461761, 0.15033191442489624, -0.0669175386428833, 0.1130417212843895, -0.010904243215918541, 0.11466218531131744, 0.026658598333597183, -0.04733501002192497, 0.1134062111377716, 0.07894708961248398, 0.07203119248151779, -0.04883268475532532, 0.00704802805557847, 0.04000711441040039, -0.0718294233083725, 0.04804560914635658, -0.045518677681684494, 0.06608755886554718, -0.13197503983974457, -0.015915190801024437, -0.08285263180732727, 0.041232481598854065, -0.04472446069121361, -0.0463087297976017, -0.04052373021841049, 0.05215632542967796, 0.07131075859069824, -0.04498962312936783, 0.06816667318344116, 0.011393166147172451, 0.07813622802495956, 0.06972888857126236, 0.08447423577308655, -0.04550296068191528, -0.10351104289293289, -0.00773888174444437, -0.008621376007795334, 0.03336777165532112, -0.11817649751901627, 0.02052859403192997, 0.1363627165555954, 0.05046108365058899, 0.12565580010414124, 0.033507172018289566, -0.04095769673585892, -0.016447054222226143, 0.032948222011327744, -0.12804651260375977, -0.09502853453159332, 0.029538221657276154, -0.07703699916601181, -0.10702905058860779, 0.014645003713667393, 0.09761476516723633, -0.03593587875366211, -0.02031797170639038, 0.004600783344358206, 0.03158603236079216, 0.008914030157029629, 0.22233085334300995, 0.024218201637268066, 0.06358204782009125, -0.11432700604200363, 0.12510792911052704, 0.0638497844338417, -0.08992574363946915, 0.04642368108034134, 0.12787672877311707, -0.09797827154397964, -0.012581996619701385, 0.10270847380161285, 0.14485427737236023, -0.027897760272026062, -0.02998896688222885, -0.09063996374607086, -0.08633539080619812, 0.07871024310588837, 0.12317868322134018, 0.030919527634978294, 0.008034615777432919, -0.053759969770908356, 0.014001740142703056, -0.14940431714057922, 0.06371805816888809, 0.08939782530069351, 0.056924402713775635, -0.08412505686283112, 0.14537549018859863, 0.0257867481559515, 0.0408640094101429, -0.017612675204873085, 0.012183005921542645, -0.07030008733272552, -0.010720420628786087, -0.07440555095672607, -0.006632300093770027, -0.014085069298744202, -0.003392559476196766, -0.02088519185781479, -0.0482097826898098, -0.03666021674871445, 0.04142361879348755, -0.0707051232457161, -0.05695244297385216, -0.004068434238433838, 0.03780748322606087, -0.13464178144931793, -0.0015053261304274201, 0.0014189650537446141, -0.09298045188188553, 0.07514927536249161, 0.04883295297622681, -0.005005288869142532, 0.029182709753513336, -0.1471504271030426, -0.031636010855436325, 0.033891964703798294, 0.012254738248884678, 0.07210731506347656, -0.07948701083660126, -0.001674938714131713, -0.02567424438893795, 0.03752048686146736, 0.01817203313112259, 0.044994015246629715, -0.11556891351938248, 0.01530440803617239, -0.06535930186510086, -0.03435812518000603, -0.06890808045864105, 0.037099312990903854, 0.10495096445083618, 0.035418108105659485, 0.1776302456855774, -0.07814448326826096, 0.041718821972608566, -0.18119853734970093, -0.032046254724264145, -0.000685486535076052, -0.03908299654722214, -0.05816365405917168, -0.012847138568758965, 0.11084382236003876, -0.053591981530189514, 0.12687106430530548, 0.014176900498569012, 0.06144491583108902, 0.03676403686404228, -0.02142753079533577, -0.066007100045681, 0.01007996965199709, 0.14225633442401886, 0.05819431692361832, -0.022296585142612457, 0.11179251968860626, -0.00845941249281168, 0.049390073865652084, 0.049689386039972305, 0.22807173430919647, 0.11519584059715271, 0.02320309542119503, 0.06966246664524078, 0.05111660808324814, -0.13523992896080017, -0.12285847961902618, 0.10219413787126541, -0.08773065358400345, 0.12188206613063812, -0.06459615379571915, 0.20231057703495026, 0.05335625261068344, -0.14798739552497864, 0.057460617274045944, -0.020613281056284904, -0.10623355209827423, -0.10616082698106766, -0.023134401068091393, -0.07539556175470352, -0.10337020456790924, 0.03567218780517578, -0.11084731668233871, 0.0645160898566246, 0.09909217059612274, 0.03286071866750717, 0.01809791289269924, 0.11484190076589584, 0.0004980289377272129, -0.000807599164545536, 0.0640435665845871, 0.024914389476180077, 0.0005387516575865448, -0.06531214714050293, -0.06411273777484894, 0.0226769857108593, 0.025168782100081444, 0.08609282970428467, -0.03678709268569946, -0.0007956248591654003, 0.04276121035218239, -0.018668469041585922, -0.07300656288862228, 0.03681676834821701, 0.005476079415529966, 0.06645271927118301, 0.05672236531972885, 0.058263394981622696, 0.008047742769122124, -0.04461135342717171, 0.2783106565475464, -0.056091129779815674, -0.10592720657587051, -0.12026821076869965, 0.22314375638961792, 0.050763003528118134, -0.024671606719493866, 0.0645555853843689, -0.10951407253742218, -0.04478717967867851, 0.15271559357643127, 0.1692090481519699, -0.05249379575252533, -0.026174768805503845, -0.03373481705784798, -0.010499906726181507, -0.015608326531946659, 0.12076389044523239, 0.09962588548660278, 0.05479196086525917, -0.05210794880986214, -0.019162671640515327, -0.015004124492406845, -0.03891700133681297, -0.07750603556632996, 0.06391581147909164, 0.018882466480135918, -0.004209671635180712, -0.03341802582144737, 0.07556223124265671, -0.02318478375673294, -0.19110526144504547, 0.03523508086800575, -0.16621042788028717, -0.18469375371932983, -0.002494668588042259, 0.09073953330516815, -0.02449381910264492, 0.050022099167108536, 0.015613138675689697, -0.014294730499386787, 0.09573032706975937, -0.03175223618745804, -0.04049323499202728, -0.10495895892381668, 0.0637938603758812, -0.10995909571647644, 0.23034454882144928, -0.0018128755036741495, 0.07650754600763321, 0.0892006978392601, 0.020642006769776344, -0.12771211564540863, 0.06333121657371521, 0.040992558002471924, -0.08851427584886551, 0.029408108443021774, 0.15896634757518768, -0.04235900938510895, 0.025314562022686005, 0.033804092556238174, -0.1116223856806755, -0.00564478849992156, -0.0755539983510971, -0.0214439257979393, -0.07129312306642532, -0.016894038766622543, -0.027055788785219193, 0.14821334183216095, 0.209843710064888, -0.007428513839840889, 0.029968395829200745, -0.07865720987319946, -0.0047265407629311085, 0.023953594267368317, 0.07532328367233276, -0.03225279599428177, -0.22306160628795624, 0.033706992864608765, 0.012218987569212914, 0.03934348002076149, -0.17362792789936066, -0.0725775808095932, 0.01078445091843605, -0.07224598526954651, -0.05140048265457153, 0.10475696623325348, 0.03954683989286423, 0.04036771133542061, -0.022034499794244766, -0.08411641418933868, -0.01014230027794838, 0.1492294818162918, -0.19551943242549896, -0.04238013178110123 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta_bert_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "roberta_bert_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/roberta_bert_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# roberta_bert_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
[ "# roberta_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# roberta_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ 62, 39, 6, 12, 8, 3, 118, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# roberta_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ -0.0824057012796402, 0.13038146495819092, -0.003791416296735406, 0.06863989681005478, 0.14306706190109253, 0.03896064683794975, 0.10725846886634827, 0.10550432652235031, -0.11848334968090057, 0.05410608649253845, 0.077451191842556, 0.07272849977016449, 0.03848404809832573, 0.15991854667663574, -0.03725622966885567, -0.2099641114473343, -0.002263711765408516, -0.00850650854408741, -0.07843461632728577, 0.11907859146595001, 0.0826350525021553, -0.10629864037036896, 0.05766814947128296, -0.006764598656445742, -0.14267852902412415, 0.02185109257698059, -0.03726084157824516, -0.05173647776246071, 0.09656445682048798, -0.006247339304536581, 0.0957479402422905, 0.03513353690505028, 0.14680711925029755, -0.1960141807794571, 0.005689928308129311, 0.09777073562145233, 0.04287108778953552, 0.09355620294809341, 0.053803153336048126, -0.016293231397867203, 0.13125300407409668, -0.14950266480445862, 0.10374777764081955, 0.01544500607997179, -0.07848180830478668, -0.11181362718343735, -0.10373780876398087, 0.053800009191036224, 0.08022775501012802, 0.12483178079128265, 0.00624672370031476, 0.13398417830467224, -0.11889348179101944, 0.08243340998888016, 0.17922455072402954, -0.20124603807926178, -0.0484122671186924, 0.05744586139917374, 0.01954338140785694, 0.06570163369178772, -0.08949077129364014, -0.01819944567978382, 0.025295410305261612, 0.02251579239964485, 0.09599761664867401, 0.004518812522292137, -0.09457375109195709, -0.000009239528480975423, -0.12628315389156342, -0.019848641008138657, 0.06650904566049576, 0.023356055840849876, -0.010903277434408665, -0.0865543931722641, -0.053613170981407166, -0.14305035769939423, -0.02722233347594738, -0.0073861898854374886, 0.03676074370741844, -0.05945153906941414, -0.08137842267751694, -0.039256371557712555, -0.06295253336429596, -0.04966669902205467, -0.0314757414162159, 0.13510309159755707, 0.0462590754032135, 0.00811672955751419, -0.04889107868075371, 0.11425881832838058, 0.0779329314827919, -0.12065361440181732, 0.020966948941349983, 0.012003672309219837, -0.08691107481718063, -0.045071784406900406, -0.0307595394551754, -0.008860291913151741, -0.004023488145321608, 0.13636475801467896, -0.07827425748109818, 0.07739957422018051, 0.025340888649225235, -0.013443033210933208, -0.015881028026342392, 0.14107368886470795, -0.04855986684560776, -0.03849635273218155, -0.00702147139236331, 0.11146889626979828, 0.006723219063133001, -0.01311876904219389, -0.06442910432815552, 0.007184130605310202, 0.06574518233537674, 0.06100347638130188, -0.04006307199597359, 0.047219350934028625, -0.014378402382135391, -0.011968124657869339, 0.04285816848278046, -0.1367427259683609, 0.03957335278391838, 0.01714138314127922, -0.10012824833393097, -0.010757241398096085, 0.010456866584718227, -0.0013314972165971994, -0.04652728512883186, 0.1375984251499176, -0.07508253306150436, 0.004738567862659693, -0.08322978764772415, -0.08064236491918564, 0.006774160545319319, -0.11010941118001938, -0.035311535000801086, -0.038738131523132324, -0.16394184529781342, -0.06225575506687164, 0.02190183475613594, -0.053574543446302414, -0.04095647856593132, -0.05739109218120575, -0.0760931670665741, 0.020170651376247406, -0.012611244805157185, 0.19139760732650757, -0.0696038231253624, 0.07297034561634064, -0.02683805674314499, 0.02562631107866764, 0.05712612718343735, 0.04440506920218468, -0.06656722724437714, -0.01551514957100153, -0.10055519640445709, 0.08491022139787674, -0.08039217442274094, 0.009726939722895622, -0.10486665368080139, -0.09536775946617126, 0.023796046152710915, -0.006972048431634903, 0.06325525790452957, 0.12219366431236267, -0.19211170077323914, -0.03436030074954033, 0.11017800122499466, -0.04951247572898865, -0.0049826898612082005, 0.04993109777569771, -0.05864083766937256, -0.007789678871631622, 0.06447219848632812, 0.12894873321056366, 0.06042627990245819, -0.12340225279331207, -0.002029975177720189, -0.0035388024989515543, 0.031681593507528305, 0.007336168084293604, 0.012841551564633846, 0.009721783921122551, 0.058313414454460144, 0.006732387933880091, -0.038468990474939346, 0.013433254323899746, -0.10327953100204468, -0.06328340619802475, -0.03366530314087868, -0.07979482412338257, 0.006384117528796196, 0.02828953042626381, 0.03249288350343704, -0.06349862366914749, -0.1124800369143486, 0.07527820020914078, 0.12677118182182312, -0.054272446781396866, 0.020031947642564774, -0.07746013253927231, -0.03359030559659004, 0.016852665692567825, -0.020915905013680458, -0.2032684087753296, -0.10268860310316086, 0.017787931486964226, -0.04086124151945114, 0.04714319109916687, -0.01883694715797901, 0.059537485241889954, 0.03770333155989647, -0.03463836386799812, -0.01525792945176363, -0.06079286336898804, 0.0023214910179376602, -0.09208999574184418, -0.21612298488616943, -0.027682466432452202, -0.015323128551244736, 0.16329146921634674, -0.2146790772676468, 0.010636366903781891, -0.024468418210744858, 0.14248935878276825, 0.023389872163534164, -0.056877583265304565, 0.007453002966940403, 0.06023651361465454, 0.010207059793174267, -0.09100429713726044, 0.03518123924732208, -0.00405983068048954, -0.07359203696250916, -0.011711612343788147, -0.1407901495695114, -0.020706666633486748, 0.0866098627448082, 0.07767243683338165, -0.1076442301273346, 0.027386032044887543, -0.07483033835887909, -0.03607316315174103, -0.08734487742185593, 0.04721243679523468, 0.2127595990896225, 0.03245236724615097, 0.1326722502708435, -0.049716413021087646, -0.08462908118963242, -0.007662066724151373, 0.022326089441776276, 0.019964637234807014, 0.11354745179414749, 0.08911622315645218, -0.04704276844859123, 0.06766913831233978, 0.014138412661850452, -0.04446638748049736, 0.12346699833869934, -0.04292794689536095, -0.08347638696432114, 0.007728120777755976, -0.035817865282297134, -0.025464843958616257, 0.09123171120882034, -0.02607494592666626, 0.007762714754790068, 0.029565323144197464, 0.027504781261086464, 0.03493589162826538, -0.18916383385658264, -0.0011350669665262103, 0.012953132390975952, -0.0535152330994606, -0.04828663170337677, -0.01648210920393467, 0.05773364007472992, 0.10709671676158905, 0.0071805366314947605, -0.0320853590965271, 0.003461911343038082, -0.0054116821847856045, -0.06892519444227219, 0.1710314154624939, -0.11191818863153458, -0.1570221185684204, -0.06851854175329208, 0.004651162773370743, -0.0212851669639349, -0.05582786723971367, 0.0077712866477668285, -0.12106364965438843, -0.069869764149189, -0.11475040018558502, -0.023602772504091263, 0.007969058118760586, -0.0004331406962592155, 0.04998626187443733, 0.01801747828722, 0.044752806425094604, -0.13730601966381073, 0.01557142473757267, -0.06415806710720062, -0.07870647311210632, 0.02204502746462822, 0.08363641053438187, 0.062470756471157074, 0.147843137383461, -0.019594712182879448, 0.028793146833777428, -0.01911352016031742, 0.20716062188148499, -0.09182767570018768, 0.006787472870200872, 0.10531491041183472, 0.017827073112130165, 0.04209775850176811, 0.12026186287403107, 0.04069599509239197, -0.0788436010479927, 0.03336925059556961, 0.09625592082738876, -0.02219819836318493, -0.2607719898223877, -0.06465498358011246, -0.016260137781500816, -0.09564731270074844, 0.08128555864095688, 0.05532560124993324, -0.024515386670827866, 0.0070046246983110905, -0.005765452049672604, -0.015345324762165546, -0.006756103131920099, 0.0495547279715538, 0.07682334631681442, 0.05700284615159035, 0.09464596956968307, -0.035517338663339615, -0.02630782127380371, 0.06439586728811264, -0.013034859672188759, 0.22701744735240936, -0.07458119839429855, 0.05665149912238121, 0.029107291251420975, 0.0888759195804596, -0.006989671383053064, 0.02458254247903824, 0.019237197935581207, -0.013977323658764362, 0.010186619125306606, -0.04481329768896103, 0.00012987013906240463, -0.011762806214392185, -0.017118623480200768, -0.005229067988693714, -0.06756193190813065, 0.03851815685629845, 0.012566015124320984, 0.2729944884777069, 0.02936982922255993, -0.2825126647949219, -0.049123365432024, -0.028155101463198662, -0.031498827040195465, -0.04093436896800995, -0.006566288415342569, 0.08417028188705444, -0.10872279107570648, 0.06085610389709473, -0.03067694790661335, 0.08755585551261902, -0.07234886288642883, 0.0013892925344407558, 0.06189108267426491, 0.1480208933353424, -0.019266698509454727, 0.054600790143013, -0.21586966514587402, 0.22315526008605957, 0.003747117705643177, 0.1257539689540863, -0.05862861126661301, 0.011675490066409111, 0.015883520245552063, 0.03761293366551399, 0.05400584265589714, -0.005612234119325876, -0.01748432219028473, -0.15853464603424072, -0.11357204616069794, 0.030515285208821297, 0.12049589306116104, 0.00794284325093031, 0.08151304721832275, -0.025068452581763268, -0.013988376595079899, 0.03931088745594025, -0.16832122206687927, -0.1600557416677475, -0.12410913407802582, 0.024392390623688698, 0.035436853766441345, -0.047278858721256256, -0.05435435473918915, -0.11937500536441803, -0.03606310486793518, 0.17012077569961548, 0.03192923218011856, -0.05519673600792885, -0.1423124223947525, 0.06797044724225998, 0.15004929900169373, -0.04313196241855621, 0.010863551869988441, 0.03488042205572128, 0.10717267543077469, 0.04804285243153572, -0.08338619023561478, 0.03267574682831764, -0.05926040560007095, -0.18822047114372253, -0.05292893573641777, 0.12310321629047394, 0.08840322494506836, 0.05104467272758484, -0.02577025443315506, 0.027603980153799057, -0.002328477567061782, -0.08068890124559402, 0.002663570223376155, 0.0821286290884018, 0.0785716250538826, 0.06194281578063965, -0.06581704318523407, -0.006367148365825415, -0.049385763704776764, -0.028588656336069107, 0.10723129659891129, 0.20292574167251587, -0.07875677943229675, 0.10652319341897964, 0.06935805082321167, -0.06414227932691574, -0.180455282330513, 0.05605823174118996, 0.12693031132221222, 0.007540388498455286, 0.024768995121121407, -0.19641079008579254, 0.1218859925866127, 0.11729106307029724, -0.008569619618356228, 0.043206438422203064, -0.3433314859867096, -0.12705688178539276, 0.0530315563082695, 0.09212003648281097, 0.03244654834270477, -0.09841720759868622, -0.014111452735960484, -0.017830124124884605, -0.1421651989221573, 0.13823513686656952, -0.056854236871004105, 0.11126962304115295, -0.00515920901671052, 0.10314076393842697, 0.02004593051970005, -0.043026067316532135, 0.09968972951173782, 0.056962158530950546, 0.07227390259504318, -0.04592418298125267, 0.009249129332602024, 0.02835470251739025, -0.06053842604160309, 0.024279089644551277, -0.04204831272363663, 0.07656276226043701, -0.10762431472539902, -0.013830609619617462, -0.08485867828130722, 0.03413385897874832, -0.04392879456281662, -0.04225188493728638, -0.041866645216941833, 0.04948088899254799, 0.07559133321046829, -0.04728953540325165, 0.0580340176820755, -0.0034254551865160465, 0.08049021661281586, 0.046880707144737244, 0.08853188902139664, -0.061201225966215134, -0.07808312773704529, -0.0054976800456643105, -0.012980479747056961, 0.030979502946138382, -0.11554446816444397, 0.021242184564471245, 0.13277602195739746, 0.05095137283205986, 0.11315611749887466, 0.03907380253076553, -0.03388892486691475, -0.01631842739880085, 0.03618590906262398, -0.12013117969036102, -0.07692456990480423, 0.03617751970887184, -0.059105049818754196, -0.1003628671169281, 0.01318759098649025, 0.10459281504154205, -0.03171771392226219, -0.023850739002227783, 0.00002300314918102231, 0.024681024253368378, 0.015499095432460308, 0.2139028012752533, 0.031107835471630096, 0.061273232102394104, -0.12105067819356918, 0.12409000843763351, 0.0648416131734848, -0.08794288337230682, 0.04638077691197395, 0.13515488803386688, -0.10581428557634354, -0.00984592642635107, 0.10295474529266357, 0.14375297725200653, -0.033184122294187546, -0.027432911098003387, -0.09412994980812073, -0.09716897457838058, 0.08638350665569305, 0.14220640063285828, 0.030401846393942833, -0.0016386250499635935, -0.053943291306495667, 0.01531407330185175, -0.15538105368614197, 0.060772575438022614, 0.08505172282457352, 0.055180780589580536, -0.07951843738555908, 0.13467592000961304, 0.030569816008210182, 0.03244614973664284, -0.017913006246089935, 0.014235163107514381, -0.07668617367744446, -0.0051995255053043365, -0.0733879879117012, -0.009534566663205624, -0.0024808086454868317, -0.0005001245881430805, -0.024806281551718712, -0.05269177258014679, -0.04579858109354973, 0.04410814866423607, -0.07601723819971085, -0.05624381825327873, -0.000047592995542800054, 0.036516495048999786, -0.135270893573761, 0.006465531419962645, 0.002306153066456318, -0.09133830666542053, 0.07120887190103531, 0.04655630514025688, -0.008410231210291386, 0.03324488550424576, -0.1696689873933792, -0.05085669085383415, 0.028277942910790443, 0.01582481898367405, 0.08183379471302032, -0.07404246926307678, -0.00007089194696163759, -0.01731390506029129, 0.046465300023555756, 0.014565679244697094, 0.05598611384630203, -0.11383845657110214, 0.015003127045929432, -0.06785861402750015, -0.03433361276984215, -0.07216466963291168, 0.039262063801288605, 0.10933477431535721, 0.046092890202999115, 0.17373016476631165, -0.08033466339111328, 0.03260502964258194, -0.1787988245487213, -0.0317988246679306, -0.007235207129269838, -0.04558084160089493, -0.08038419485092163, -0.012241165153682232, 0.11487498879432678, -0.05008617043495178, 0.12626874446868896, 0.01269222516566515, 0.061090435832738876, 0.04099923372268677, -0.03080131486058235, -0.06890372931957245, 0.011931156739592552, 0.14700397849082947, 0.06317096203565598, -0.017590470612049103, 0.1144009456038475, -0.011856756173074245, 0.0482829287648201, 0.050796277821063995, 0.22918008267879486, 0.14000022411346436, 0.04855858162045479, 0.06320314854383469, 0.04329865425825119, -0.13772401213645935, -0.09864413738250732, 0.1144774854183197, -0.09992320090532303, 0.11792457848787308, -0.06841877847909927, 0.21159981191158295, 0.057062529027462006, -0.13557536900043488, 0.05121072754263878, -0.03209617733955383, -0.0972064808011055, -0.10636674612760544, 0.004399748984724283, -0.07918056845664978, -0.107582226395607, 0.041173044592142105, -0.11119262874126434, 0.06585895270109177, 0.09834190458059311, 0.031102396547794342, 0.030806010589003563, 0.09570225328207016, -0.0027024478185921907, 0.001425672904588282, 0.05646849051117897, 0.020471248775720596, -0.007158348802477121, -0.0649489015340805, -0.06676706671714783, 0.03077545575797558, 0.02552873082458973, 0.08613363653421402, -0.032745543867349625, -0.014618810266256332, 0.049536339938640594, -0.01920623891055584, -0.07219056040048599, 0.04300585389137268, 0.0027127163484692574, 0.07463470846414566, 0.06101970374584198, 0.052029069513082504, 0.00929994136095047, -0.04705897346138954, 0.2942902147769928, -0.05802009627223015, -0.12645146250724792, -0.11823051422834396, 0.24295227229595184, 0.05880143865942955, -0.02619738131761551, 0.06227663904428482, -0.10105184465646744, -0.049350544810295105, 0.16902144253253937, 0.17844878137111664, -0.06726489961147308, -0.026139922440052032, -0.029980219900608063, -0.011630825698375702, -0.031641580164432526, 0.13637536764144897, 0.12230902165174484, 0.05470471829175949, -0.04968297854065895, -0.02722051925957203, -0.02562405727803707, -0.03300796449184418, -0.06770487874746323, 0.06914003938436508, 0.02274605818092823, -0.007867845706641674, -0.03898024931550026, 0.06896993517875671, -0.014606191776692867, -0.20764607191085815, 0.030621228739619255, -0.16617995500564575, -0.18565593659877777, -0.0053040022030472755, 0.09942243993282318, -0.027191802859306335, 0.04573539271950722, 0.016587864607572556, -0.012039384804666042, 0.09030033648014069, -0.03417803347110748, -0.02552429400384426, -0.09912370145320892, 0.062391072511672974, -0.12668053805828094, 0.22394460439682007, -0.01040998101234436, 0.07381571829319, 0.09499037265777588, 0.029608409851789474, -0.12546788156032562, 0.07213129848241806, 0.041574202477931976, -0.10058601200580597, 0.03471206873655319, 0.15569280087947845, -0.042833372950553894, 0.03948650509119034, 0.039448101073503494, -0.11891087889671326, 0.004116350784897804, -0.09822811186313629, -0.030271684750914574, -0.060943786054849625, -0.020428761839866638, -0.0303350742906332, 0.14757533371448517, 0.2093755304813385, -0.007797469850629568, 0.0329802967607975, -0.08312954753637314, -0.010553035885095596, 0.019558914005756378, 0.08138524740934372, -0.029212648048996925, -0.22848956286907196, 0.03477820008993149, 0.019498905166983604, 0.0341450497508049, -0.19077612459659576, -0.06270970404148102, 0.00022777076810598373, -0.0704704001545906, -0.054039083421230316, 0.1099073737859726, 0.04452647641301155, 0.04151095449924469, -0.024022391065955162, -0.08641368895769119, -0.009721397422254086, 0.15065714716911316, -0.19051365554332733, -0.038572024554014206 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta_distilgpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.2 - Tokenizers 0.11.0
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "roberta_distilgpt2_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/roberta_distilgpt2_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# roberta_distilgpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.2 - Tokenizers 0.11.0
[ "# roberta_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.2\n- Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# roberta_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.2\n- Tokenizers 0.11.0" ]
[ 62, 43, 6, 12, 8, 3, 118, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# roberta_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.2\n- Tokenizers 0.11.0" ]
[ -0.08227404206991196, 0.14844098687171936, -0.003707609139382839, 0.06674090027809143, 0.13195085525512695, 0.04939545318484306, 0.11033397167921066, 0.12197372317314148, -0.10854723304510117, 0.05316091701388359, 0.08603111654520035, 0.04770134389400482, 0.05075396969914436, 0.15477825701236725, -0.03361516445875168, -0.22620654106140137, 0.005524700973182917, -0.006776683963835239, -0.08109670132398605, 0.12314728647470474, 0.08746571838855743, -0.09639442712068558, 0.05706559494137764, -0.010560333728790283, -0.13933336734771729, 0.02513265423476696, -0.04645795375108719, -0.05958431959152222, 0.09370165318250656, -0.01070401631295681, 0.09260066598653793, 0.04064098745584488, 0.15101195871829987, -0.20955075323581696, 0.001735592377372086, 0.08606931567192078, 0.042011938989162445, 0.09201835095882416, 0.058812834322452545, -0.014592879451811314, 0.1177569180727005, -0.1685320883989334, 0.09554393589496613, 0.011613648384809494, -0.07253335416316986, -0.11467266827821732, -0.09879451245069504, 0.08478163182735443, 0.08575313538312912, 0.11721257865428925, 0.0021691820584237576, 0.12980499863624573, -0.11638879776000977, 0.07791951298713684, 0.16112664341926575, -0.20188668370246887, -0.05298880115151405, 0.0592927411198616, 0.016898801550269127, 0.06412097811698914, -0.08671645075082779, -0.015340938232839108, 0.030989516526460648, 0.018841248005628586, 0.08861128240823746, 0.006121708080172539, -0.08908721059560776, 0.005607615225017071, -0.12296249717473984, -0.036049410700798035, 0.11842404305934906, 0.03148239850997925, -0.010621733032166958, -0.0999986007809639, -0.042411573231220245, -0.1392805576324463, -0.01340471301227808, -0.01329907588660717, 0.03434683755040169, -0.06534846127033234, -0.07197706401348114, -0.05141512304544449, -0.07655731588602066, -0.048709575086832047, -0.011913901194930077, 0.10290082544088364, 0.044138792902231216, 0.010654509998857975, -0.04547810181975365, 0.12291885167360306, 0.07343164086341858, -0.12562507390975952, 0.0034430602099746466, 0.009455434046685696, -0.08817534148693085, -0.051847897469997406, -0.024149255827069283, -0.005823361687362194, -0.005299829877912998, 0.134439617395401, -0.06324649602174759, 0.07226479798555374, 0.02280903421342373, -0.010703605599701405, -0.015672411769628525, 0.13564617931842804, -0.04849061369895935, -0.04693170636892319, -0.021180909126996994, 0.11508246511220932, 0.0009773168712854385, -0.005315360147505999, -0.068873830139637, -0.005818863864988089, 0.07774774730205536, 0.059554897248744965, -0.0385298877954483, 0.04827377572655678, -0.015223645605146885, -0.021190887317061424, 0.03556016832590103, -0.14475640654563904, 0.03663259744644165, 0.01942036673426628, -0.09726130217313766, -0.006082931533455849, 0.019668810069561005, -0.004020061809569597, -0.05179719999432564, 0.1184602603316307, -0.07450096309185028, 0.006656738929450512, -0.07883206754922867, -0.06756269931793213, 0.018021246418356895, -0.10505304485559464, -0.034631695598363876, -0.03494269400835037, -0.1888401359319687, -0.06157273054122925, 0.02307097055017948, -0.05554661899805069, -0.05249777436256409, -0.06178797036409378, -0.08189678937196732, 0.01578126661479473, -0.011105608195066452, 0.18546800315380096, -0.05989552289247513, 0.07118198275566101, -0.019725555554032326, 0.021421926096081734, 0.06884727627038956, 0.04335243999958038, -0.06587570905685425, -0.004331981297582388, -0.08963858336210251, 0.09265341609716415, -0.08295278996229172, 0.002582057612016797, -0.10706286877393723, -0.10415783524513245, 0.01228325441479683, -0.010600132867693901, 0.043911293148994446, 0.14391639828681946, -0.1787458211183548, -0.03614399582147598, 0.1342441588640213, -0.05739365145564079, -0.006558106280863285, 0.06398212164640427, -0.05561084672808647, -0.004309221636503935, 0.05688444897532463, 0.13127437233924866, 0.07347487658262253, -0.11776712536811829, -0.010655234567821026, 0.00302622327581048, 0.03893861174583435, 0.008046893402934074, 0.03182145208120346, 0.00982636772096157, 0.05749800428748131, 0.008627360686659813, -0.06313305348157883, 0.0065926616080105305, -0.10148782283067703, -0.06967149674892426, -0.03818998485803604, -0.07860817760229111, 0.03395902365446091, 0.030908789485692978, 0.043063148856163025, -0.05720606818795204, -0.11457518488168716, 0.08982601016759872, 0.12826058268547058, -0.05817783251404762, 0.003706812858581543, -0.08019081503152847, -0.04225568845868111, 0.008083237335085869, -0.022168399766087532, -0.19379660487174988, -0.11128085851669312, 0.018373144790530205, -0.04047265648841858, 0.04675944149494171, -0.003314064349979162, 0.06739974766969681, 0.03184426948428154, -0.04478466883301735, -0.024744153022766113, -0.07422720640897751, 0.0006941421306692064, -0.08212131261825562, -0.19787068665027618, -0.04673989117145538, -0.01996970921754837, 0.19294221699237823, -0.2338264435529709, 0.024794047698378563, -0.013336540199816227, 0.1499091237783432, 0.02088610641658306, -0.05843843147158623, 0.019299428910017014, 0.051257163286209106, 0.005996576976031065, -0.08940143138170242, 0.025524301454424858, -0.009595165029168129, -0.0796421691775322, -0.014505266211926937, -0.1300407499074936, -0.024638814851641655, 0.07609499990940094, 0.09098023176193237, -0.10978849232196808, 0.01812836527824402, -0.07825342565774918, -0.0476134829223156, -0.08868690580129623, 0.02271343395113945, 0.19720081984996796, 0.03434937074780464, 0.12954187393188477, -0.0456375926733017, -0.08828361332416534, -0.0022878912277519703, 0.012550696730613708, 0.0003706142888404429, 0.10722909867763519, 0.07411109656095505, -0.06438018381595612, 0.08073005825281143, 0.015914563089609146, -0.04921938106417656, 0.1445659101009369, -0.052881013602018356, -0.09462225437164307, -0.0010076836915686727, -0.018305707722902298, -0.028994515538215637, 0.09972148388624191, -0.051920127123594284, -0.007521770428866148, 0.03495357930660248, 0.021754637360572815, 0.041160836815834045, -0.17992348968982697, -0.0052805934101343155, 0.01540153007954359, -0.05897364392876625, -0.038578201085329056, -0.02851966582238674, 0.05861306190490723, 0.10240566730499268, 0.0022475013975054026, -0.03356558084487915, 0.021106550469994545, -0.007351531647145748, -0.07913246005773544, 0.1721048355102539, -0.10592914372682571, -0.1740138679742813, -0.10020243376493454, 0.04139845818281174, -0.027123097330331802, -0.054221730679273605, 0.0004855705192312598, -0.11221626400947571, -0.05945910885930061, -0.11293892562389374, -0.023961365222930908, -0.0074758767150342464, -0.011263005435466766, 0.058933984488248825, 0.024156516417860985, 0.04982605576515198, -0.12429331988096237, 0.017919206991791725, -0.043426819145679474, -0.07908514887094498, 0.029504623264074326, 0.05840775743126869, 0.06126302480697632, 0.1440034955739975, -0.021077148616313934, 0.03440725803375244, -0.019543005153536797, 0.18035922944545746, -0.09787657856941223, 0.01011285837739706, 0.10984702408313751, 0.013780334033071995, 0.0379016250371933, 0.11191374808549881, 0.03258652612566948, -0.06484351307153702, 0.02303447760641575, 0.08760269731283188, -0.02195708639919758, -0.26555806398391724, -0.06438974291086197, -0.014360531233251095, -0.07202865928411484, 0.08914440870285034, 0.05893220752477646, -0.006833454594016075, 0.012455510906875134, -0.011633004993200302, -0.018528982996940613, -0.00033680754131637514, 0.05383049696683884, 0.08154263347387314, 0.043975960463285446, 0.08870315551757812, -0.03580417111515999, -0.032753147184848785, 0.0625033900141716, 0.0030395458452403545, 0.23095864057540894, -0.05866042152047157, 0.07328952848911285, 0.01393992267549038, 0.10274601727724075, -0.010816818103194237, 0.02825554646551609, 0.03721674159169197, -0.007499521132558584, 0.016108276322484016, -0.042436473071575165, -0.004537700675427914, 0.005091853439807892, -0.004778399132192135, -0.005136588588356972, -0.06485392898321152, 0.04340449348092079, 0.007756212260574102, 0.2904805839061737, 0.04426248371601105, -0.2910598814487457, -0.05802645534276962, -0.019857419654726982, -0.038686588406562805, -0.048297666013240814, -0.011792842298746109, 0.10309024900197983, -0.1238236203789711, 0.06832118332386017, -0.03959399461746216, 0.08150053024291992, -0.0683000311255455, -0.00007890418055467308, 0.05177908390760422, 0.11600136011838913, -0.010772895999252796, 0.07309772819280624, -0.20809277892112732, 0.23343084752559662, 0.0041764527559280396, 0.11743086576461792, -0.070938341319561, 0.02170003205537796, 0.010515639558434486, 0.0458606518805027, 0.08155764639377594, -0.007184295449405909, -0.015466786921024323, -0.14795167744159698, -0.1291855424642563, 0.027347005903720856, 0.11834266781806946, -0.01611548289656639, 0.08476146310567856, -0.019285494461655617, -0.01520137395709753, 0.03257715702056885, -0.15477479994297028, -0.15255044400691986, -0.1294572949409485, 0.03569972142577171, 0.03110172040760517, -0.03402199223637581, -0.05725495517253876, -0.11863460391759872, -0.031005989760160446, 0.1976701319217682, 0.0065496861934661865, -0.058440741151571274, -0.14562192559242249, 0.0799887552857399, 0.1510588675737381, -0.03846625238656998, -0.00026565263397060335, 0.04362005740404129, 0.11846311390399933, 0.044318776577711105, -0.08178582787513733, 0.022950751706957817, -0.05920768156647682, -0.19186565279960632, -0.044436633586883545, 0.13537432253360748, 0.08144698292016983, 0.05468236282467842, -0.014579528011381626, 0.023194793611764908, 0.005538728553801775, -0.07273760437965393, 0.009924469515681267, 0.08001775294542313, 0.08719678968191147, 0.06995894014835358, -0.0719456672668457, -0.011022781953215599, -0.048451557755470276, -0.028457103297114372, 0.1096482053399086, 0.18161506950855255, -0.08372855931520462, 0.09909919649362564, 0.06156705319881439, -0.0690988153219223, -0.1767812818288803, 0.046840135008096695, 0.13458116352558136, 0.0196326095610857, 0.029990818351507187, -0.2063002586364746, 0.11468806862831116, 0.12841932475566864, -0.010814039036631584, 0.027918951585888863, -0.34068772196769714, -0.11813995242118835, 0.05939583480358124, 0.07100578397512436, 0.017423558980226517, -0.1061769351363182, -0.02607368491590023, 0.0016989829018712044, -0.14013445377349854, 0.13328048586845398, -0.05700237303972244, 0.10409331321716309, -0.00478620920330286, 0.11890842020511627, 0.027562618255615234, -0.0507475845515728, 0.11936777085065842, 0.07395677268505096, 0.06131725385785103, -0.04632466658949852, 0.0069138421677052975, 0.05607076734304428, -0.06963890790939331, 0.05692873150110245, -0.04104525223374367, 0.06920056790113449, -0.13944111764431, -0.019311701878905296, -0.0879695788025856, 0.04373077675700188, -0.043677590787410736, -0.03775694966316223, -0.028592849150300026, 0.05464714765548706, 0.06506729125976562, -0.04200389236211777, 0.07116756588220596, -0.0006597075262106955, 0.08297254890203476, 0.08039523661136627, 0.10384180396795273, -0.027963971719145775, -0.08627587556838989, -0.02168714441359043, -0.016324391588568687, 0.03270256146788597, -0.10535623878240585, 0.018821345642209053, 0.1355435848236084, 0.05640783905982971, 0.13311856985092163, 0.030000271275639534, -0.0382896289229393, -0.018391352146863937, 0.03789949789643288, -0.12529514729976654, -0.10762036591768265, 0.015074729919433594, -0.06589405238628387, -0.12154531478881836, 0.01170484721660614, 0.10659963637590408, -0.034705083817243576, -0.02439974807202816, -0.002214649925008416, 0.03224336355924606, 0.008218524977564812, 0.21070973575115204, 0.025137560442090034, 0.07280313968658447, -0.11122887581586838, 0.11664829403162003, 0.06690996140241623, -0.10150644183158875, 0.05062583461403847, 0.12465313822031021, -0.09948751330375671, -0.017353922128677368, 0.08989163488149643, 0.13188065588474274, -0.03309003263711929, -0.03677116706967354, -0.09495925903320312, -0.09929589182138443, 0.07885729521512985, 0.11583898961544037, 0.0338728167116642, 0.010060589760541916, -0.04396182298660278, 0.006548203062266111, -0.1529051661491394, 0.06786943227052689, 0.09327539801597595, 0.054799165576696396, -0.09447141736745834, 0.14863668382167816, 0.017941167578101158, 0.044381849467754364, -0.01668532006442547, 0.013388944789767265, -0.06361044198274612, -0.006534905638545752, -0.07469990104436874, -0.005165495909750462, -0.006987604312598705, -0.004299296997487545, -0.013808023184537888, -0.048341862857341766, -0.037865038961172104, 0.0416739359498024, -0.07983393222093582, -0.05548863112926483, -0.0019945718813687563, 0.046224601566791534, -0.12987884879112244, 0.002872479148209095, 0.004421238321810961, -0.09889189898967743, 0.076186403632164, 0.05658983439207077, -0.004707470536231995, 0.02976212278008461, -0.1297268569469452, -0.037414707243442535, 0.03157948702573776, 0.022188544273376465, 0.06693922728300095, -0.06808946281671524, 0.002726944861933589, -0.015330314636230469, 0.03277861326932907, 0.024573106318712234, 0.04392484948039055, -0.11163073033094406, 0.009738816879689693, -0.06903353333473206, -0.03593333810567856, -0.07818283885717392, 0.04898860305547714, 0.10626300424337387, 0.0420684777200222, 0.1667836308479309, -0.08068504929542542, 0.03655752167105675, -0.1858992874622345, -0.027844350785017014, -0.00378011935390532, -0.03166225925087929, -0.059750497341156006, -0.012095196172595024, 0.11083492636680603, -0.045956648886203766, 0.11889328062534332, 0.003707622177898884, 0.05503560230135918, 0.03860995173454285, -0.04356842115521431, -0.05954359099268913, 0.01585240662097931, 0.15228089690208435, 0.06092699617147446, -0.020005203783512115, 0.10786899924278259, -0.025788499042391777, 0.033147308975458145, 0.04546694830060005, 0.21886608004570007, 0.14043524861335754, 0.011266660876572132, 0.06870956718921661, 0.060697879642248154, -0.13755296170711517, -0.10908307135105133, 0.11037742346525192, -0.08707892149686813, 0.12319963425397873, -0.06556084752082825, 0.2055101990699768, 0.0677976980805397, -0.14627081155776978, 0.04808703437447548, -0.03258177638053894, -0.10611869394779205, -0.10273667424917221, -0.019386785104870796, -0.08222021907567978, -0.09890325367450714, 0.04184066504240036, -0.1168922483921051, 0.05931074917316437, 0.09498272091150284, 0.03418401628732681, 0.022640347480773926, 0.10067474097013474, -0.0018508655484765768, -0.0037201757077127695, 0.06900390982627869, 0.01688457652926445, -0.0022103101946413517, -0.08565793931484222, -0.06080286204814911, 0.04184066504240036, 0.03190618008375168, 0.08850201964378357, -0.03820857033133507, 0.003823197679594159, 0.03569386526942253, -0.018656963482499123, -0.07424186915159225, 0.034826673567295074, 0.0022607301361858845, 0.07282042503356934, 0.057364340871572495, 0.04887082800269127, -0.0010150335729122162, -0.05347006395459175, 0.2773582935333252, -0.06427310407161713, -0.11706894636154175, -0.1389189213514328, 0.22776980698108673, 0.043565139174461365, -0.025408143177628517, 0.07025279849767685, -0.10763293504714966, -0.033398665487766266, 0.152565598487854, 0.17675691843032837, -0.03900834545493126, -0.023668216541409492, -0.026510437950491905, -0.01072661206126213, -0.02278320863842964, 0.11196853965520859, 0.10113167017698288, 0.05871990695595741, -0.051973726600408554, -0.021251648664474487, -0.010741982609033585, -0.03395325690507889, -0.07471026480197906, 0.07457150518894196, 0.019319772720336914, -0.008614144288003445, -0.03901347517967224, 0.07750679552555084, -0.013768468983471394, -0.19864541292190552, 0.043877579271793365, -0.17324085533618927, -0.1860542744398117, -0.00907705444842577, 0.09168293327093124, -0.027283091098070145, 0.046854641288518906, 0.013398214243352413, -0.020221585407853127, 0.10788802802562714, -0.031556256115436554, -0.048570968210697174, -0.09039776027202606, 0.06657703220844269, -0.1140017881989479, 0.2241080105304718, -0.0059847342781722546, 0.0808219239115715, 0.08964729309082031, 0.012374348938465118, -0.13083532452583313, 0.05278228595852852, 0.041872162371873856, -0.07010859996080399, 0.033747170120477676, 0.17331674695014954, -0.044513385742902756, 0.04584071412682533, 0.03310451656579971, -0.12766852974891663, -0.016122419387102127, -0.08128703385591507, -0.026053527370095253, -0.06567154824733734, -0.01230720803141594, -0.021769601851701736, 0.1476389616727829, 0.2156156748533249, -0.015017551369965076, 0.026971258223056793, -0.07871625572443008, -0.0029136857483536005, 0.029081720858812332, 0.08176810294389725, -0.01892048679292202, -0.21773368120193481, 0.023708375170826912, -0.018490925431251526, 0.036332085728645325, -0.20046581327915192, -0.07246921211481094, 0.004167203791439533, -0.06247615069150925, -0.048731349408626556, 0.10609707981348038, 0.06416530907154083, 0.04412413761019707, -0.019017623737454414, -0.0612037368118763, -0.01229055505245924, 0.14642666280269623, -0.19130265712738037, -0.03415582701563835 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta_gpt2_new_max64_summarization_cnndm This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "roberta_gpt2_new_max64_summarization_cnndm", "results": []}]}
text2text-generation
Ayham/roberta_gpt2_new_max64_summarization_cnndm
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# roberta_gpt2_new_max64_summarization_cnndm This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
[ "# roberta_gpt2_new_max64_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# roberta_gpt2_new_max64_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ 62, 44, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# roberta_gpt2_new_max64_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ -0.08914168924093246, 0.15646275877952576, -0.0034035572316497564, 0.07504424452781677, 0.13263387978076935, 0.044081538915634155, 0.09954015165567398, 0.1443473994731903, -0.10727700591087341, 0.07229094952344894, 0.08783864974975586, 0.02925124578177929, 0.053603824228048325, 0.1518949717283249, -0.03258892893791199, -0.24240362644195557, 0.010897734202444553, -0.004014178179204464, -0.06928183883428574, 0.1104133129119873, 0.08778435736894608, -0.09482338279485703, 0.07314952462911606, -0.013476599007844925, -0.13056498765945435, 0.014404207468032837, -0.03507049009203911, -0.06073623523116112, 0.08814164251089096, -0.017896147444844246, 0.09148687869310379, 0.01930968090891838, 0.1420053243637085, -0.20387738943099976, -0.0009019823046401143, 0.08168166875839233, 0.039849329739809036, 0.09101521968841553, 0.058513835072517395, -0.012854431755840778, 0.11647018045186996, -0.17656254768371582, 0.09223000705242157, 0.013310161419212818, -0.07779546827077866, -0.0971374660730362, -0.09224763512611389, 0.06438611447811127, 0.09118780493736267, 0.10537045449018478, 0.012527057901024818, 0.11164442449808121, -0.09104694426059723, 0.08738983422517776, 0.16512912511825562, -0.22639669477939606, -0.06343960762023926, 0.08214841783046722, 0.044718168675899506, 0.0679624080657959, -0.09718558937311172, -0.010060751810669899, 0.024536550045013428, 0.026505043730139732, 0.1025567278265953, -0.012842209078371525, -0.08499903231859207, 0.008572558872401714, -0.12487037479877472, -0.039342209696769714, 0.12887494266033173, 0.02091353014111519, -0.013365487568080425, -0.09613273292779922, -0.0596122108399868, -0.14310166239738464, -0.016560325399041176, -0.020188508555293083, 0.032578833401203156, -0.052121855318546295, -0.06398075819015503, -0.06633774936199188, -0.07321180403232574, -0.06746795028448105, -0.006379256024956703, 0.09406964480876923, 0.0348413847386837, 0.016168877482414246, -0.04557206854224205, 0.1283828169107437, 0.03940696269273758, -0.12659382820129395, -0.009935206733644009, 0.009092179127037525, -0.09991540014743805, -0.05559086799621582, -0.025852175429463387, -0.03227078914642334, -0.009337407536804676, 0.13617153465747833, -0.04898828640580177, 0.08457031846046448, 0.015294631943106651, -0.009777999483048916, -0.026976237073540688, 0.13849608600139618, -0.045956432819366455, -0.06339142471551895, -0.006489516701549292, 0.10762012749910355, 0.005718244705349207, -0.007798468694090843, -0.06474044919013977, -0.02729862928390503, 0.08171513676643372, 0.056779805570840836, -0.03771411255002022, 0.042892731726169586, -0.016225207597017288, -0.024362832307815552, 0.02644791081547737, -0.13767506182193756, 0.04226719215512276, 0.008079794235527515, -0.09156887233257294, -0.006030996795743704, 0.028732286766171455, -0.012393685057759285, -0.04280568286776543, 0.12250955402851105, -0.07888897508382797, 0.0000013835077652402106, -0.07907788455486298, -0.06181167811155319, 0.023030292242765427, -0.10295368731021881, -0.04357827827334404, -0.03131704032421112, -0.22322246432304382, -0.05519359931349754, 0.02938363328576088, -0.05535166710615158, -0.03642063960433006, -0.05991065874695778, -0.07397329807281494, 0.025280511006712914, -0.0068672229535877705, 0.1568424552679062, -0.05472629889845848, 0.07660713791847229, -0.015199883840978146, 0.027875125408172607, 0.0531780906021595, 0.038677673786878586, -0.06891558319330215, 0.009401329793035984, -0.12579719722270966, 0.0876237228512764, -0.07686641067266464, 0.007514291908591986, -0.11143691837787628, -0.10172407329082489, -0.006414579693228006, -0.016615085303783417, 0.044014230370521545, 0.1471187323331833, -0.18916481733322144, -0.023814527317881584, 0.15183432400226593, -0.06704969704151154, -0.013139206916093826, 0.06520668417215347, -0.05768651142716408, 0.006485991179943085, 0.06406079232692719, 0.13418221473693848, 0.09079333394765854, -0.1169525757431984, -0.021269522607326508, -0.006311517208814621, 0.03318969905376434, 0.0034022934269160032, 0.03923490643501282, 0.009348341263830662, 0.05725906789302826, 0.007897555828094482, -0.07688488066196442, -0.00760646304115653, -0.08290231972932816, -0.073701411485672, -0.051782652735710144, -0.07375741004943848, 0.03317391127347946, 0.020141780376434326, 0.044087037444114685, -0.0559854693710804, -0.10452386736869812, 0.09358164668083191, 0.13653114438056946, -0.06451135128736496, 0.010145227424800396, -0.07234031707048416, -0.02596728503704071, -0.008531575091183186, -0.0225094985216856, -0.19680482149124146, -0.10532931983470917, 0.032618921250104904, -0.038151174783706665, 0.046700965613126755, 0.00007094792817952111, 0.0759451612830162, 0.040262382477521896, -0.04796324297785759, -0.02725258097052574, -0.06312961876392365, -0.009523610584437847, -0.0891844853758812, -0.17835655808448792, -0.04291893541812897, -0.022615399211645126, 0.16296547651290894, -0.23302292823791504, 0.026658453047275543, 0.009480779990553856, 0.1588476300239563, 0.031583450734615326, -0.06282150000333786, 0.025565266609191895, 0.04418877884745598, 0.007893114350736141, -0.09762544184923172, 0.023750465363264084, -0.008104678243398666, -0.0781233161687851, -0.03912418335676193, -0.13237176835536957, -0.002009442076086998, 0.06642238050699234, 0.09304223209619522, -0.09993010759353638, 0.006789178587496281, -0.05890299752354622, -0.057599931955337524, -0.08923817425966263, 0.00023161459830589592, 0.19110067188739777, 0.03727518022060394, 0.1277334988117218, -0.05063634365797043, -0.0719086155295372, -0.001988997682929039, 0.028146419674158096, 0.004557985812425613, 0.08447589725255966, 0.06272469460964203, -0.07399679720401764, 0.07636494934558868, 0.018184078857302666, -0.06075308471918106, 0.1506119966506958, -0.049172282218933105, -0.09040424227714539, -0.017084946855902672, 0.000778061046730727, -0.01790851540863514, 0.10399110615253448, -0.04651039093732834, -0.010780430398881435, 0.028928400948643684, 0.026190832257270813, 0.03219931572675705, -0.17063914239406586, 0.00040874414844438434, 0.01406783889979124, -0.07001183182001114, -0.0071872929111123085, -0.00725294416770339, 0.04357977956533432, 0.0900442823767662, 0.005344073753803968, -0.03270987421274185, 0.019626136869192123, -0.01061219722032547, -0.07910376787185669, 0.17216654121875763, -0.10715948045253754, -0.1695919632911682, -0.1237642914056778, 0.05038870498538017, -0.037503186613321304, -0.04134350270032883, -0.003957369364798069, -0.08321645855903625, -0.05467478930950165, -0.10257457196712494, -0.04233846440911293, -0.024868162348866463, -0.00461775716394186, 0.034481555223464966, 0.011514399200677872, 0.057494841516017914, -0.12879395484924316, 0.014858290553092957, -0.023628482595086098, -0.0817374438047409, 0.016325775533914566, 0.055411551147699356, 0.07056760042905807, 0.13014063239097595, -0.018658116459846497, 0.0322544127702713, -0.02696429006755352, 0.18361742794513702, -0.08794917166233063, 0.022407256066799164, 0.11671802401542664, 0.006280084606260061, 0.05173966661095619, 0.10835093259811401, 0.02810288779437542, -0.06792347878217697, 0.02093355916440487, 0.06993064284324646, -0.021246444433927536, -0.2536812722682953, -0.05341293290257454, -0.02617243491113186, -0.04731297492980957, 0.10094370692968369, 0.06146526336669922, -0.0009103927295655012, 0.022961875423789024, -0.009019318968057632, 0.00974089652299881, -0.012294609099626541, 0.0744561031460762, 0.09272828698158264, 0.04758027568459511, 0.09353143721818924, -0.033688995987176895, -0.027370795607566833, 0.0631604790687561, 0.010677771642804146, 0.24034246802330017, -0.04193972051143646, 0.09723047912120819, 0.004292139317840338, 0.10659697651863098, -0.011823480017483234, 0.04114300012588501, 0.02504694275557995, 0.006888324860483408, 0.011224225163459778, -0.056432563811540604, -0.018863312900066376, 0.019065091386437416, -0.017009638249874115, 0.008213185705244541, -0.06556673347949982, 0.03675342723727226, 0.013347186148166656, 0.289751797914505, 0.03979948163032532, -0.2972813844680786, -0.05604514107108116, -0.0027554200496524572, -0.04118357226252556, -0.05940697714686394, -0.0064982143230736256, 0.12420099973678589, -0.13570162653923035, 0.069381482899189, -0.06494058668613434, 0.083646260201931, -0.08078892529010773, -0.0005972839426249266, 0.05791899189352989, 0.11718921363353729, -0.011106791906058788, 0.06799823045730591, -0.20349912345409393, 0.21281935274600983, 0.01875079795718193, 0.11725292354822159, -0.06729190796613693, 0.026145849376916885, 0.0221767146140337, 0.0479113832116127, 0.08889681845903397, -0.009815320372581482, -0.06137116998434067, -0.13672396540641785, -0.1321164071559906, 0.027052033692598343, 0.12061747908592224, -0.02384876273572445, 0.08523933589458466, -0.02524224855005741, -0.016280489042401314, 0.03275018557906151, -0.14551515877246857, -0.14796775579452515, -0.11006053537130356, 0.037995845079422, 0.04663195461034775, -0.018736470490694046, -0.06128992140293121, -0.10376418381929398, -0.01188466977328062, 0.17955715954303741, -0.020811209455132484, -0.05051009729504585, -0.14260071516036987, 0.05622711405158043, 0.15525364875793457, -0.04839906468987465, 0.018689053133130074, 0.03598775342106819, 0.11669570952653885, 0.042096707969903946, -0.07764748483896255, 0.025146160274744034, -0.05801215022802353, -0.1795855313539505, -0.05469222366809845, 0.13931117951869965, 0.0671989917755127, 0.048417966812849045, -0.012651204131543636, 0.032591987401247025, 0.006702973507344723, -0.08476083725690842, -0.004583615809679031, 0.08928114920854568, 0.0996224582195282, 0.0708489939570427, -0.07352882623672485, 0.007514442317187786, -0.03615434840321541, -0.01624438911676407, 0.1178915947675705, 0.1634618490934372, -0.08187076449394226, 0.09333618730306625, 0.04683256894350052, -0.07307323068380356, -0.17228199541568756, 0.05903596058487892, 0.11750522255897522, 0.02975505217909813, 0.04074818640947342, -0.20690913498401642, 0.1002582311630249, 0.11940737068653107, -0.01908748783171177, 0.04971921443939209, -0.3333089351654053, -0.11748329550027847, 0.041479527950286865, 0.07472848892211914, 0.015335586853325367, -0.11666285246610641, -0.03921050578355789, -0.0011489969911053777, -0.12926144897937775, 0.12876847386360168, -0.07197795808315277, 0.11116643249988556, -0.016945799812674522, 0.11885559558868408, 0.02757779136300087, -0.052774105221033096, 0.12465095520019531, 0.06321565806865692, 0.06773586571216583, -0.04050936549901962, 0.006726834457367659, 0.06455832719802856, -0.08118025213479996, 0.06751298904418945, -0.06336697936058044, 0.06884323060512543, -0.1659889668226242, -0.008877006359398365, -0.07653221487998962, 0.04450615122914314, -0.045096226036548615, -0.03985603153705597, -0.04164263606071472, 0.05371078848838806, 0.06079401075839996, -0.03085331618785858, 0.06514973938465118, 0.0089953588321805, 0.07231814414262772, 0.10305588692426682, 0.09900519251823425, -0.0003121359331998974, -0.09999246895313263, -0.012743963859975338, -0.010878259316086769, 0.0405430793762207, -0.1296500712633133, 0.018632955849170685, 0.14106231927871704, 0.04546371474862099, 0.1401616334915161, 0.0293082557618618, -0.0405140146613121, -0.017461854964494705, 0.043383482843637466, -0.13391335308551788, -0.09637103974819183, 0.009710822254419327, -0.06265497207641602, -0.13215294480323792, 0.010524868033826351, 0.11705470085144043, -0.04552854225039482, -0.014535117894411087, 0.0034004016779363155, 0.030042745172977448, -0.0042836833745241165, 0.20741799473762512, 0.020028332248330116, 0.06710970401763916, -0.09912431985139847, 0.11597811430692673, 0.07249412685632706, -0.08195433765649796, 0.0458361990749836, 0.11926603317260742, -0.09666631370782852, -0.01064988225698471, 0.0652477890253067, 0.14648810029029846, -0.03134290501475334, -0.03311000391840935, -0.08691388368606567, -0.08127478510141373, 0.0650157704949379, 0.11939506977796555, 0.03237432613968849, 0.017370741814374924, -0.05726407840847969, 0.016077116131782532, -0.138450488448143, 0.0702754408121109, 0.09001640975475311, 0.06440828740596771, -0.11424872279167175, 0.1535453349351883, 0.0059691425412893295, 0.03256578370928764, -0.014399497769773006, 0.01563745178282261, -0.06657548248767853, -0.015681419521570206, -0.07073725014925003, 0.0019870870746672153, -0.01743422821164131, -0.009810741990804672, -0.015662137418985367, -0.04332922771573067, -0.043769195675849915, 0.03350742533802986, -0.06890131533145905, -0.05543544888496399, -0.00861743837594986, 0.04606820270419121, -0.1286986619234085, -0.003919824492186308, 0.010292897932231426, -0.09417995810508728, 0.07447301596403122, 0.05767381936311722, 0.00470268027856946, 0.02699066698551178, -0.09736005961894989, -0.016449877992272377, 0.033253204077482224, 0.019847586750984192, 0.053283028304576874, -0.06831535696983337, 0.0014680195599794388, -0.026211077347397804, 0.03798649087548256, 0.026228489354252815, 0.04041457548737526, -0.11296022683382034, 0.011017235927283764, -0.06917253136634827, -0.02199738472700119, -0.07043153047561646, 0.04213220253586769, 0.12096784263849258, 0.040979646146297455, 0.1646578162908554, -0.07786940038204193, 0.039358753710985184, -0.1937015801668167, -0.01903381571173668, -0.001458632294088602, -0.04710112139582634, -0.06656216830015182, -0.019116761162877083, 0.10222325474023819, -0.05196439474821091, 0.10890425741672516, -0.005936237517744303, 0.056129902601242065, 0.03511867672204971, -0.03910325840115547, -0.0602245070040226, 0.008284814655780792, 0.14721384644508362, 0.06946099549531937, -0.026585988700389862, 0.11242711544036865, -0.01843416690826416, 0.04314330592751503, 0.049124788492918015, 0.20874916017055511, 0.12958259880542755, 0.0027548731304705143, 0.08087233453989029, 0.058369092643260956, -0.1117902398109436, -0.12705515325069427, 0.11533141881227493, -0.059556882828474045, 0.11673486977815628, -0.057510267943143845, 0.18753810226917267, 0.07499274611473083, -0.16318634152412415, 0.04575834423303604, -0.029639262706041336, -0.11118672788143158, -0.10455320030450821, -0.04195312410593033, -0.08068426698446274, -0.09535922855138779, 0.02671463042497635, -0.11736395955085754, 0.05898444354534149, 0.07609675824642181, 0.025660816580057144, 0.01278284378349781, 0.12155614793300629, -0.023697003722190857, 0.004560164175927639, 0.07370243966579437, 0.023915214464068413, 0.011537059210240841, -0.050857674330472946, -0.07146701961755753, 0.026390094310045242, 0.028286701068282127, 0.09241613745689392, -0.04132264852523804, 0.02336047776043415, 0.034144558012485504, -0.015079386532306671, -0.06852007657289505, 0.025402752682566643, 0.016296448186039925, 0.05584415793418884, 0.06517922133207321, 0.05914856493473053, -0.0067772711627185345, -0.04560166224837303, 0.2718321681022644, -0.0748104378581047, -0.08885549753904343, -0.1305261105298996, 0.19189327955245972, 0.04356991872191429, -0.022393539547920227, 0.060249220579862595, -0.10704072564840317, -0.028164783492684364, 0.15700975060462952, 0.15714287757873535, -0.051441486924886703, -0.017677608877420425, -0.02162017673254013, -0.011513063684105873, -0.017314812168478966, 0.09817787259817123, 0.0756610706448555, 0.05918356776237488, -0.04737675189971924, -0.01710759848356247, -0.0024935004767030478, -0.03624647855758667, -0.07963676750659943, 0.04782295227050781, -0.001333705149590969, -0.0004508255224209279, -0.022554628551006317, 0.07564011961221695, -0.02232309617102146, -0.18882055580615997, 0.03426363691687584, -0.1770596206188202, -0.18255376815795898, -0.0038265949115157127, 0.07466930150985718, -0.023645464330911636, 0.041374221444129944, -0.0008423480321653187, -0.01687891036272049, 0.10254333913326263, -0.028913849964737892, -0.05153365805745125, -0.09961462765932083, 0.06911706179380417, -0.09142210334539413, 0.21263542771339417, 0.0010766901541501284, 0.07185579091310501, 0.08802080154418945, 0.014700017869472504, -0.12987250089645386, 0.03981834277510643, 0.048606544733047485, -0.07986999303102493, 0.025480609387159348, 0.17933900654315948, -0.05057721957564354, 0.06569258123636246, 0.02943411096930504, -0.1063016727566719, -0.021998325362801552, -0.03371533751487732, -0.012641227804124355, -0.07900645583868027, -0.023335905745625496, -0.032991182059049606, 0.1626567244529724, 0.2133907526731491, -0.014175496995449066, 0.023301323875784874, -0.07054467499256134, 0.009350713342428207, 0.03436518833041191, 0.0778360664844513, -0.026297561824321747, -0.21497106552124023, 0.016811750829219818, 0.004217572044581175, 0.03548332676291466, -0.19210168719291687, -0.09049441665410995, 0.021044185385107994, -0.05180581659078598, -0.0526200532913208, 0.1126878634095192, 0.04584486037492752, 0.033289916813373566, -0.014541826210916042, -0.081329345703125, -0.01942630484700203, 0.1447547823190689, -0.19200070202350616, -0.047725506126880646 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta_gpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description This model uses RoBerta encoder and GPT2 decoder and fine-tuned on the summarization task. It got Rouge scores as follows: Rouge1= 35.886 Rouge2= 16.292 RougeL= 23.499 ## Intended uses & limitations To use its API: from transformers import RobertaTokenizerFast, GPT2Tokenizer, EncoderDecoderModel model = EncoderDecoderModel.from_pretrained("Ayham/roberta_gpt2_summarization_cnn_dailymail") input_tokenizer = RobertaTokenizerFast.from_pretrained('roberta-base') output_tokenizer = GPT2Tokenizer.from_pretrained("gpt2") article = """Your Input Text""" input_ids = input_tokenizer(article, return_tensors="pt").input_ids output_ids = model.generate(input_ids) print(output_tokenizer.decode(output_ids[0], skip_special_tokens=True)) More information needed More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "roberta_gpt2_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/roberta_gpt2_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# roberta_gpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description This model uses RoBerta encoder and GPT2 decoder and fine-tuned on the summarization task. It got Rouge scores as follows: Rouge1= 35.886 Rouge2= 16.292 RougeL= 23.499 ## Intended uses & limitations To use its API: from transformers import RobertaTokenizerFast, GPT2Tokenizer, EncoderDecoderModel model = EncoderDecoderModel.from_pretrained("Ayham/roberta_gpt2_summarization_cnn_dailymail") input_tokenizer = RobertaTokenizerFast.from_pretrained('roberta-base') output_tokenizer = GPT2Tokenizer.from_pretrained("gpt2") article = """Your Input Text""" input_ids = input_tokenizer(article, return_tensors="pt").input_ids output_ids = model.generate(input_ids) print(output_tokenizer.decode(output_ids[0], skip_special_tokens=True)) More information needed More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# roberta_gpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\nThis model uses RoBerta encoder and GPT2 decoder and fine-tuned on the summarization task. It got Rouge scores as follows:\n\nRouge1= 35.886\n\nRouge2= 16.292\n\nRougeL= 23.499", "## Intended uses & limitations\nTo use its API:\n\nfrom transformers import RobertaTokenizerFast, GPT2Tokenizer, EncoderDecoderModel\n\nmodel = EncoderDecoderModel.from_pretrained(\"Ayham/roberta_gpt2_summarization_cnn_dailymail\")\n\ninput_tokenizer = RobertaTokenizerFast.from_pretrained('roberta-base')\n\noutput_tokenizer = GPT2Tokenizer.from_pretrained(\"gpt2\")\n\narticle = \"\"\"Your Input Text\"\"\"\n\ninput_ids = input_tokenizer(article, return_tensors=\"pt\").input_ids\n\noutput_ids = model.generate(input_ids)\n\nprint(output_tokenizer.decode(output_ids[0], skip_special_tokens=True))\n\nMore information needed\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# roberta_gpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\nThis model uses RoBerta encoder and GPT2 decoder and fine-tuned on the summarization task. It got Rouge scores as follows:\n\nRouge1= 35.886\n\nRouge2= 16.292\n\nRougeL= 23.499", "## Intended uses & limitations\nTo use its API:\n\nfrom transformers import RobertaTokenizerFast, GPT2Tokenizer, EncoderDecoderModel\n\nmodel = EncoderDecoderModel.from_pretrained(\"Ayham/roberta_gpt2_summarization_cnn_dailymail\")\n\ninput_tokenizer = RobertaTokenizerFast.from_pretrained('roberta-base')\n\noutput_tokenizer = GPT2Tokenizer.from_pretrained(\"gpt2\")\n\narticle = \"\"\"Your Input Text\"\"\"\n\ninput_ids = input_tokenizer(article, return_tensors=\"pt\").input_ids\n\noutput_ids = model.generate(input_ids)\n\nprint(output_tokenizer.decode(output_ids[0], skip_special_tokens=True))\n\nMore information needed\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 62, 41, 59, 216, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# roberta_gpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\nThis model uses RoBerta encoder and GPT2 decoder and fine-tuned on the summarization task. It got Rouge scores as follows:\n\nRouge1= 35.886\n\nRouge2= 16.292\n\nRougeL= 23.499## Intended uses & limitations\nTo use its API:\n\nfrom transformers import RobertaTokenizerFast, GPT2Tokenizer, EncoderDecoderModel\n\nmodel = EncoderDecoderModel.from_pretrained(\"Ayham/roberta_gpt2_summarization_cnn_dailymail\")\n\ninput_tokenizer = RobertaTokenizerFast.from_pretrained('roberta-base')\n\noutput_tokenizer = GPT2Tokenizer.from_pretrained(\"gpt2\")\n\narticle = \"\"\"Your Input Text\"\"\"\n\ninput_ids = input_tokenizer(article, return_tensors=\"pt\").input_ids\n\noutput_ids = model.generate(input_ids)\n\nprint(output_tokenizer.decode(output_ids[0], skip_special_tokens=True))\n\nMore information needed\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results" ]
[ -0.04625774919986725, 0.2024948000907898, -0.008103407919406891, 0.07383020967245102, 0.08960241079330444, 0.0326739177107811, 0.05842467024922371, 0.15010268986225128, -0.0599864162504673, 0.14351509511470795, 0.0833924189209938, 0.09804289042949677, 0.06858402490615845, 0.09337957948446274, 0.013847604393959045, -0.15384916961193085, 0.029706565663218498, -0.06556009501218796, -0.05750495195388794, 0.08341024816036224, 0.08455561846494675, -0.049485236406326294, 0.0871802493929863, 0.04089857265353203, -0.041382480412721634, 0.014351124875247478, -0.03413258120417595, -0.040568575263023376, 0.03513147681951523, 0.04856275022029877, 0.04342233017086983, 0.00008621811866760254, 0.09977968782186508, -0.22425563633441925, 0.006321428343653679, 0.10211780667304993, -0.004006603267043829, 0.08746699243783951, 0.15674792230129242, -0.01989651471376419, 0.10313922166824341, -0.13823235034942627, 0.08814103156328201, 0.06036924198269844, -0.1201917976140976, -0.10535404086112976, -0.15046189725399017, 0.0979539304971695, 0.08492939919233322, 0.0965634435415268, -0.01799309067428112, 0.07625420391559601, -0.049620043486356735, 0.06082037091255188, 0.13749295473098755, -0.18233788013458252, -0.043788254261016846, 0.04220915213227272, 0.049377456307411194, -0.04949929192662239, -0.0913139134645462, -0.014577822759747505, -0.0017649655928835273, -0.0014062327099964023, 0.025318119674921036, -0.010161100886762142, 0.06253043562173843, 0.005657551810145378, -0.11320231109857559, -0.060560744255781174, 0.0733637735247612, 0.07730799168348312, -0.03580743074417114, -0.1378016620874405, -0.01807645708322525, -0.09848467260599136, -0.012754429131746292, -0.03697968274354935, 0.006725176237523556, -0.010807695798575878, -0.03730791062116623, -0.026734577491879463, -0.04657285660505295, -0.028150513768196106, -0.010592648759484291, 0.0033150408416986465, 0.023865966126322746, -0.010618191212415695, -0.010228829458355904, 0.09503212571144104, -0.0021140603348612785, -0.1002979427576065, -0.03735671564936638, -0.020489294081926346, -0.12431371957063675, -0.03956976532936096, 0.0109616843983531, -0.006288424134254456, 0.01669248566031456, 0.23947718739509583, -0.004224047996103764, 0.09316448867321014, -0.0019356400007382035, -0.017518073320388794, 0.007728312164545059, 0.16796986758708954, -0.10110849142074585, -0.08188522607088089, -0.04516904056072235, 0.055822111666202545, -0.0390746109187603, -0.014619187451899052, -0.01136874221265316, 0.00469158124178648, 0.03033522330224514, 0.031078699976205826, 0.06247050315141678, 0.0340726301074028, -0.05057825520634651, -0.03757255896925926, 0.12002141773700714, -0.12361092120409012, 0.059658050537109375, 0.04382285103201866, -0.09848286956548691, 0.045601628720760345, 0.05771080031991005, -0.01198914647102356, -0.0911126658320427, 0.10216362774372101, -0.07567676156759262, -0.014042454771697521, -0.03068077377974987, -0.07863595336675644, 0.003115120343863964, -0.026524266228079796, -0.0607423260807991, -0.028948403894901276, -0.09484986215829849, -0.09370497614145279, 0.05922288820147514, -0.08389322459697723, -0.0076277959160506725, -0.05911706015467644, -0.059503283351659775, 0.07351026684045792, 0.041211601346731186, 0.03263691067695618, -0.0308174230158329, 0.020561547949910164, -0.05318307504057884, 0.015589072369039059, 0.12261950224637985, 0.027231700718402863, -0.02882096916437149, 0.04390100762248039, -0.07768689841032028, 0.12616291642189026, -0.06174182519316673, -0.005435305181890726, -0.1595938503742218, -0.07418733090162277, -0.0008044937276281416, -0.009412657469511032, 0.05707293376326561, 0.11149163544178009, -0.12197469174861908, -0.027941958978772163, 0.15134231746196747, -0.02447720803320408, -0.01848919317126274, 0.0676439180970192, -0.060445886105298996, -0.013408217579126358, 0.04208144173026085, 0.10169242322444916, 0.08036098629236221, -0.13413697481155396, -0.054845020174980164, -0.02472628653049469, -0.02563338913023472, 0.12792877852916718, 0.0594756044447422, -0.09816782176494598, 0.03401835262775421, -0.00041571364272385836, -0.040794480592012405, -0.005799585487693548, -0.04438285902142525, -0.04323848336935043, -0.0033964733593165874, -0.003965360578149557, 0.02482384257018566, -0.02378104254603386, -0.0011216681450605392, -0.017726842314004898, -0.1653350591659546, -0.024481453001499176, 0.10608092695474625, -0.02425062097609043, 0.018490685150027275, -0.09302370250225067, -0.02117426134645939, 0.05288440361618996, 0.0023394597228616476, -0.1308058202266693, -0.060269106179475784, 0.011189120821654797, -0.16227391362190247, 0.05100790783762932, -0.08804303407669067, 0.052201077342033386, 0.04210183024406433, 0.0001304522156715393, -0.03245960548520088, -0.02782895229756832, -0.0265093632042408, -0.02362905628979206, -0.14596454799175262, -0.03948849067091942, 0.00802564900368452, 0.17086392641067505, -0.17302167415618896, 0.032035138458013535, 0.021133629605174065, 0.12737269699573517, 0.018486086279153824, -0.08129163086414337, 0.02174476347863674, -0.012365754693746567, 0.0008204769110307097, -0.08046518266201019, -0.011334484443068504, -0.006264638155698776, -0.06279855221509933, 0.017689943313598633, -0.16151517629623413, -0.12053889036178589, 0.06369681656360626, 0.1047658696770668, -0.1419738084077835, 0.03302307054400444, -0.01753855124115944, -0.025034597143530846, -0.043365463614463806, -0.017025474458932877, 0.12597548961639404, 0.056228674948215485, 0.10312165319919586, -0.01912880688905716, -0.08500592410564423, 0.005039437208324671, 0.004931134171783924, -0.015749359503388405, 0.10508732497692108, 0.04485062509775162, -0.10408945381641388, 0.03526762127876282, 0.0019061805214732885, 0.012744484469294548, 0.10412436723709106, 0.0075540849938988686, -0.06869839131832123, -0.07680203765630722, 0.025524111464619637, 0.03681923821568489, 0.013755912892520428, 0.013886266387999058, 0.04855721443891525, 0.0460948571562767, 0.03138549253344536, -0.001999265281483531, -0.09433629363775253, 0.030742499977350235, 0.03415568917989731, -0.028449364006519318, -0.016266781836748123, 0.02906833402812481, 0.025118988007307053, 0.04779227077960968, 0.05408332496881485, 0.040134113281965256, -0.01096750795841217, -0.03686768561601639, -0.07155556231737137, 0.1450415849685669, -0.10420241951942444, -0.13852828741073608, -0.1613583266735077, -0.061969004571437836, -0.04649755358695984, -0.021658804267644882, 0.020768191665410995, -0.09436625242233276, -0.09006199985742569, -0.10574933886528015, -0.008660990744829178, -0.011302781291306019, -0.04313026741147041, 0.02802879735827446, 0.02205941453576088, 0.0757463276386261, -0.14573509991168976, 0.016113389283418655, -0.019158653914928436, -0.06707563251256943, -0.012917947955429554, 0.06135993450880051, 0.06853640824556351, 0.12853917479515076, 0.0495946891605854, 0.029849709942936897, 0.015493420884013176, 0.23920416831970215, -0.10531570017337799, 0.041871778666973114, 0.09083057194948196, 0.0227617546916008, 0.09121828526258469, 0.12227161228656769, 0.04189698398113251, -0.03347915783524513, 0.025707555934786797, 0.0932837650179863, 0.003794812597334385, -0.2488606721162796, -0.0475023090839386, -0.01976884715259075, -0.026272732764482498, 0.13038167357444763, 0.05833813175559044, 0.023802945390343666, 0.015059704892337322, -0.032572273164987564, 0.011334055103361607, 0.045988354831933975, 0.09450185298919678, 0.0626169741153717, 0.06743531674146652, 0.05650612711906433, -0.02218976989388466, -0.018665064126253128, 0.09005000442266464, -0.035896528512239456, 0.1449442058801651, -0.028031233698129654, 0.1502530425786972, 0.0342189185321331, 0.09199507534503937, -0.02190067432820797, 0.0004799201851710677, 0.034930430352687836, 0.026099689304828644, 0.01598471589386463, -0.058671873062849045, 0.013117660768330097, 0.02328483760356903, 0.039106328040361404, -0.013218126259744167, -0.029952077195048332, 0.022847548127174377, 0.06398032605648041, 0.17988142371177673, 0.05762770026922226, -0.23022545874118805, -0.025290990248322487, -0.007681554649025202, -0.016801301389932632, -0.06701039522886276, -0.036559443920850754, 0.013580095022916794, -0.1564701348543167, 0.056920137256383896, -0.03177059069275856, 0.09556527435779572, -0.06894537061452866, -0.011891535483300686, 0.05586397275328636, 0.11647353321313858, -0.019421443343162537, 0.042278438806533813, -0.07868783175945282, 0.13604675233364105, -0.005797529127448797, 0.07763252407312393, -0.007452375255525112, 0.07657159119844437, 0.04428994283080101, -0.020712969824671745, 0.12142539769411087, 0.006704739760607481, -0.12491697818040848, -0.14564168453216553, -0.10473539680242538, -0.03447749838232994, 0.1152716800570488, -0.06965399533510208, 0.09891648590564728, -0.03913721814751625, -0.05100540071725845, -0.011476884596049786, -0.0767969861626625, -0.13345842063426971, -0.15850956737995148, 0.07619629055261612, -0.0712631419301033, 0.054364003241062164, -0.02957466058433056, -0.037004150450229645, -0.051484860479831696, 0.1954232007265091, -0.12403138726949692, -0.02232523076236248, -0.12321580201387405, -0.009222794324159622, 0.15667478740215302, -0.0758323073387146, 0.050425101071596146, -0.0055404589511454105, 0.11683323234319687, 0.031585272401571274, -0.04439033195376396, 0.027414821088314056, -0.041838422417640686, -0.1417868584394455, -0.05525534972548485, 0.09634148329496384, 0.062198638916015625, 0.007238728925585747, -0.0013818673323839903, 0.049235787242650986, -0.026328038424253464, -0.10906572639942169, 0.0331478975713253, 0.06204095482826233, 0.018721118569374084, 0.06083781272172928, -0.011824792250990868, -0.06504904478788376, -0.08618137240409851, -0.00038508870056830347, 0.04919302836060524, 0.2187129706144333, -0.08527851104736328, 0.08705363422632217, 0.027372734621167183, -0.08255539834499359, -0.140219584107399, -0.03002362698316574, 0.09341245889663696, -0.029992422088980675, 0.051157861948013306, -0.15391428768634796, 0.08080479502677917, 0.11944212019443512, -0.017295818775892258, -0.002016763435676694, -0.3213942050933838, -0.1293150931596756, 0.030946480110287666, 0.009119220077991486, -0.10357335954904556, -0.11805442720651627, -0.07488181442022324, -0.039448440074920654, -0.16916894912719727, 0.09530536085367203, 0.003327754558995366, 0.062369514256715775, -0.002129852306097746, 0.01136984582990408, 0.035736795514822006, -0.031072987243533134, 0.14517265558242798, 0.05868009477853775, 0.027894670143723488, -0.06556796282529831, 0.005957417190074921, 0.05167526751756668, -0.09403491020202637, 0.06419654935598373, -0.0036558194551616907, 0.0682138204574585, -0.1469617784023285, 0.0035959244705736637, -0.04081754758954048, 0.021101519465446472, -0.0592123381793499, -0.009399504400789738, -0.04890349134802818, 0.06280314922332764, 0.10928387939929962, -0.011926265433430672, 0.04431949555873871, 0.007915062829852104, 0.04457540437579155, 0.11577332764863968, 0.02079070918262005, 0.11033054441213608, -0.16598960757255554, 0.0024556785356253386, 0.0020260573364794254, 0.01359771192073822, -0.10436287522315979, 0.03882893919944763, 0.08664323389530182, 0.029628846794366837, 0.13444824516773224, 0.009571307338774204, -0.06956009566783905, 0.006198844872415066, 0.05766725167632103, -0.11086255311965942, -0.08952071517705917, 0.008535885252058506, -0.03536278381943703, -0.09775286167860031, -0.0734868273139, 0.12163792550563812, 0.05931277573108673, -0.03568892180919647, 0.033790115267038345, 0.06652874499559402, 0.005400595720857382, 0.1491202712059021, 0.01478867419064045, 0.06073341891169548, -0.0648503378033638, 0.08628945797681808, 0.12587226927280426, -0.06391356885433197, 0.04836063086986542, 0.15494486689567566, -0.05416829511523247, -0.02906128764152527, 0.011059879325330257, 0.11728474497795105, -0.02830345742404461, 0.015123099088668823, -0.020013336092233658, -0.028389953076839447, 0.07139747589826584, 0.013931888155639172, 0.009870044887065887, 0.013861393555998802, -0.016199497506022453, -0.01857665367424488, -0.09586295485496521, 0.06656008213758469, 0.13385316729545593, 0.021987348794937134, -0.06233203411102295, 0.05445746332406998, 0.013692548498511314, 0.023464299738407135, 0.023693351075053215, -0.02922142669558525, -0.12204740941524506, -0.021396778523921967, -0.029291916638612747, 0.04572106897830963, -0.06160001829266548, 0.008341321721673012, -0.02105974592268467, 0.0010141649981960654, -0.03130356967449188, 0.01752380281686783, -0.05888912454247475, -0.12451229244470596, -0.0172062199562788, 0.10518132895231247, -0.12830094993114471, 0.012846263125538826, 0.07119474560022354, -0.10885176807641983, 0.04888856038451195, 0.020622655749320984, 0.008069233037531376, 0.0029046835843473673, -0.1275290995836258, -0.04584885388612747, -0.05157230421900749, -0.005305072292685509, 0.0034258845262229443, -0.15360529720783234, 0.019956007599830627, -0.07032082229852676, -0.014379782602190971, -0.014257055707275867, -0.01621498353779316, -0.12688443064689636, 0.026468336582183838, -0.041041843593120575, -0.028215285390615463, -0.07913969457149506, 0.04842676222324371, 0.0849650427699089, 0.017304394394159317, 0.13360536098480225, -0.05646759644150734, 0.059551164507865906, -0.17068925499916077, -0.022908609360456467, 0.024047207087278366, -0.01446904893964529, -0.001178230857476592, -0.019147252663969994, 0.07708623260259628, -0.022571295499801636, 0.0740845799446106, -0.008795931935310364, 0.009623232297599316, 0.030689500272274017, -0.025593934580683708, -0.06845887750387192, 0.010345226153731346, 0.07568857818841934, 0.032349925488233566, -0.02202831208705902, 0.0583166778087616, -0.0435318760573864, -0.007671018596738577, 0.03067951090633869, 0.10945916920900345, 0.12599672377109528, 0.053656406700611115, 0.028612015768885612, 0.03512231260538101, -0.10724861174821854, -0.08820661902427673, 0.12326537817716599, -0.09780757874250412, 0.08420941233634949, -0.06020955368876457, 0.06891976296901703, 0.05746214836835861, -0.14987759292125702, 0.07534299045801163, -0.00045098832924850285, -0.08000682294368744, -0.09724811464548111, -0.10036863386631012, -0.07912230491638184, -0.07995913177728653, 0.0373772531747818, -0.08912354707717896, 0.1195228174328804, 0.08291909843683243, 0.02774083986878395, 0.023950757458806038, 0.04149600490927696, -0.08313378691673279, -0.06684385240077972, 0.026883695274591446, 0.009230603463947773, -0.029339756816625595, -0.017195826396346092, 0.0033327778801321983, 0.02105170674622059, 0.03411633148789406, 0.10702101141214371, 0.01699068583548069, 0.0355776771903038, 0.03257416933774948, -0.04796413704752922, -0.07501767575740814, 0.03192570433020592, 0.016132313758134842, -0.009418844245374203, 0.09781136363744736, 0.0679784044623375, -0.0019059608457610011, -0.031612519174814224, 0.19788379967212677, -0.04095695540308952, -0.05997232720255852, -0.16787561774253845, 0.1276787519454956, 0.06962999701499939, 0.003761663567274809, 0.011526552960276604, -0.12409728765487671, -0.012998349964618683, 0.13539613783359528, 0.11719517409801483, 0.004115943796932697, -0.02449854463338852, 0.007923669181764126, 0.00285023869946599, 0.02244473248720169, 0.05021606758236885, 0.04707645624876022, 0.0791909396648407, -0.04000188410282135, 0.04770311340689659, 0.00619410490617156, -0.041300173848867416, -0.04978891834616661, 0.08822008222341537, -0.005890928208827972, 0.01844792068004608, -0.03536779433488846, 0.09482556581497192, -0.051143988966941833, -0.19916315376758575, 0.05527633801102638, -0.08258035033941269, -0.145662322640419, 0.008444170467555523, 0.04934743046760559, -0.005781419575214386, 0.08528250455856323, 0.02825765497982502, -0.05697585269808769, 0.1491164118051529, 0.0004182084812782705, -0.052847519516944885, -0.09858570992946625, 0.010616145096719265, -0.0889187604188919, 0.26547589898109436, 0.03628052771091461, 0.03012225031852722, 0.12363913655281067, -0.008428538218140602, -0.1290201097726822, 0.03124045766890049, 0.045538172125816345, -0.051439203321933746, 0.04110867157578468, 0.16114728152751923, -0.02034543640911579, 0.07283709198236465, 0.047704365104436874, -0.08223660290241241, 0.021952828392386436, -0.028427552431821823, 0.012853856198489666, -0.09070088714361191, 0.0172975305467844, -0.050925400108098984, 0.1310291439294815, 0.22696691751480103, -0.020871775224804878, 0.027150625362992287, -0.03804417699575424, 0.010469266213476658, -0.013155216351151466, 0.052899979054927826, -0.03009607456624508, -0.14091819524765015, 0.0783333033323288, -0.025877870619297028, 0.06841287016868591, -0.17023080587387085, -0.0904417559504509, 0.00475448789075017, -0.032266803085803986, 0.0017848822753876448, 0.1386924833059311, 0.042431291192770004, 0.012796357274055481, -0.012336406856775284, -0.0485600046813488, -0.01423920039087534, 0.12087904661893845, -0.09195590764284134, -0.059175219386816025 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta_gpt2_summarization_xsum This model is a fine-tuned version of [](https://huggingface.co/) on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["xsum"], "model-index": [{"name": "roberta_gpt2_summarization_xsum", "results": []}]}
text2text-generation
Ayham/roberta_gpt2_summarization_xsum
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:xsum", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us
# roberta_gpt2_summarization_xsum This model is a fine-tuned version of [](URL on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# roberta_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us \n", "# roberta_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 59, 35, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us \n# roberta_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.08549042791128159, 0.14987392723560333, -0.0024841090198606253, 0.05783962830901146, 0.1482631266117096, 0.048990730196237564, 0.10246863216161728, 0.11638082563877106, -0.09898736327886581, 0.0789363831281662, 0.09130830317735672, 0.06867953389883041, 0.060841187834739685, 0.14815667271614075, -0.035603154450654984, -0.2534312307834625, 0.019320029765367508, -0.0268210731446743, -0.09292618185281754, 0.09623316675424576, 0.08993800729513168, -0.10806610435247421, 0.07652492076158524, -0.005939848255366087, -0.1686173677444458, 0.01676025241613388, -0.04668350890278816, -0.057227034121751785, 0.09236044436693192, 0.008257318288087845, 0.09143476188182831, 0.01653212495148182, 0.12624475359916687, -0.22753824293613434, 0.004587546922266483, 0.0861409455537796, 0.0387766994535923, 0.09256735444068909, 0.060256585478782654, -0.001448673428967595, 0.1460651457309723, -0.13999465107917786, 0.1061389222741127, 0.01620623469352722, -0.07285457104444504, -0.12519162893295288, -0.08775357902050018, 0.025581471621990204, 0.07334014773368835, 0.09409588575363159, 0.004966293927282095, 0.12119725346565247, -0.09796593338251114, 0.07250307500362396, 0.16612407565116882, -0.2277740240097046, -0.06138138845562935, 0.029601100832223892, 0.0586128830909729, 0.07044917345046997, -0.10255381464958191, -0.010352203622460365, 0.0232990849763155, 0.024709947407245636, 0.08032017201185226, -0.0036801258102059364, -0.0927644819021225, 0.01433637272566557, -0.12297841161489487, -0.014972961507737637, 0.13188380002975464, 0.031084498390555382, -0.02599720098078251, -0.10295307636260986, -0.04880836606025696, -0.10679782927036285, -0.014396660961210728, -0.02857687510550022, 0.04070347175002098, -0.03850679099559784, -0.0876510739326477, -0.042753349989652634, -0.0649188682436943, -0.06366682797670364, -0.019308216869831085, 0.1519404649734497, 0.037260111421346664, 0.01324888039380312, -0.04235102981328964, 0.09592284262180328, 0.04444722831249237, -0.11364323645830154, -0.004955639597028494, -0.007591499947011471, -0.10633992403745651, -0.04594583809375763, -0.06007840856909752, -0.03219868615269661, 0.01643703691661358, 0.15157027542591095, -0.052789781242609024, 0.09967795759439468, 0.031239310279488564, -0.012187045067548752, 0.009939879179000854, 0.14273357391357422, -0.04725473374128342, -0.05543173477053642, -0.02070792391896248, 0.07601291686296463, 0.015061331912875175, -0.019771777093410492, -0.06461483240127563, -0.0034361262805759907, 0.07306753844022751, 0.05659586191177368, -0.0463329553604126, 0.0266280397772789, -0.03439578413963318, -0.025589458644390106, -0.006026260554790497, -0.12440571933984756, 0.04746649041771889, -0.0014261663891375065, -0.08228704333305359, -0.004903900902718306, 0.0010616693180054426, 0.01109133567661047, -0.02745600789785385, 0.12520523369312286, -0.08602890372276306, -0.012667196802794933, -0.08831261843442917, -0.07256466895341873, 0.021926699206233025, -0.13239355385303497, -0.00394368777051568, -0.0393994078040123, -0.16749659180641174, -0.046809811145067215, 0.06353619694709778, -0.06034841760993004, -0.02611582912504673, -0.05313880369067192, -0.03913332149386406, 0.0417470708489418, -0.00782785564661026, 0.17882274091243744, -0.05638623982667923, 0.06807210296392441, -0.029545070603489876, 0.03626411780714989, 0.011465840972959995, 0.04882829263806343, -0.0733904018998146, 0.007531681098043919, -0.10537383705377579, 0.08425921201705933, -0.06634363532066345, 0.011451470665633678, -0.11819978058338165, -0.07954838871955872, -0.025559742003679276, -0.0021847165189683437, 0.07450389862060547, 0.122395358979702, -0.18987105786800385, -0.03528023883700371, 0.11960966140031815, -0.07092883437871933, -0.059756312519311905, 0.06119425967335701, -0.05448843166232109, 0.01873183809220791, 0.04920035973191261, 0.16053606569766998, 0.04779469966888428, -0.10934124141931534, -0.030490251258015633, 0.007883865386247635, 0.047743719071149826, 0.015273050405085087, 0.052691057324409485, -0.011259996332228184, 0.03884387016296387, 0.012629379518330097, -0.025273077189922333, 0.011022190563380718, -0.07987531274557114, -0.07529392838478088, -0.03775125741958618, -0.0729040578007698, -0.013142355717718601, 0.02190736122429371, 0.029642311856150627, -0.06203468143939972, -0.1017962396144867, 0.10132639855146408, 0.12970305979251862, -0.07485967129468918, 0.018465403467416763, -0.06198621541261673, -0.016479037702083588, -0.01254549715667963, -0.019209373742341995, -0.20430731773376465, -0.09754015505313873, 0.030305195599794388, -0.05171862989664078, 0.03355651721358299, -0.022041458636522293, 0.0624270997941494, 0.049832407385110855, -0.03299999237060547, -0.011791463941335678, -0.0672491118311882, -0.002740890486165881, -0.09067486971616745, -0.21887052059173584, -0.03636511042714119, -0.028213193640112877, 0.18799801170825958, -0.20192566514015198, -0.007809880655258894, -0.008617698214948177, 0.14118875563144684, 0.03159499168395996, -0.07089190185070038, 0.0017033915501087904, 0.04301182180643082, -0.01033012568950653, -0.09718300402164459, 0.030814524739980698, 0.010724764317274094, -0.09970615059137344, -0.035653259605169296, -0.15210096538066864, -0.0007034587906673551, 0.09631922096014023, 0.06462381035089493, -0.08600139617919922, -0.03682543337345123, -0.0653582364320755, -0.03947962075471878, -0.07132457196712494, 0.024783719331026077, 0.18083788454532623, 0.01507064513862133, 0.11029504239559174, -0.06455475836992264, -0.07211616635322571, 0.012646099552512169, 0.029050342738628387, -0.015499193221330643, 0.09617381542921066, 0.13149823248386383, -0.10103262215852737, 0.08149745315313339, 0.07175789773464203, -0.03494512662291527, 0.15222828090190887, -0.04190642386674881, -0.08827435225248337, -0.009619015268981457, -0.005194520577788353, -0.020777104422450066, 0.11787910759449005, -0.08537232875823975, 0.016439389437437057, 0.026945937424898148, 0.03760102763772011, 0.05194231495261192, -0.17389686405658722, 0.006688377819955349, 0.020278383046388626, -0.03138373792171478, -0.04495958983898163, -0.0133237699046731, 0.024196352809667587, 0.0869748517870903, 0.026913950219750404, 0.009685940109193325, 0.013923649676144123, -0.004361032973974943, -0.07941830158233643, 0.19149631261825562, -0.1405412256717682, -0.15988965332508087, -0.08859746158123016, 0.032854966819286346, -0.03885282203555107, -0.03324677422642708, 0.00945791881531477, -0.10727129131555557, -0.06248406693339348, -0.09102780371904373, -0.022260542958974838, -0.058992624282836914, 0.013250432908535004, 0.03927098214626312, 0.017520714551210403, 0.040351320058107376, -0.13008461892604828, 0.015559489838778973, -0.05737922340631485, -0.059740498661994934, 0.0080557307228446, 0.06799479573965073, 0.09093113988637924, 0.11421767622232437, -0.010987500660121441, 0.027026288211345673, -0.028482697904109955, 0.21081407368183136, -0.08529743552207947, -0.008265350945293903, 0.10270022600889206, -0.006288711912930012, 0.04307879880070686, 0.11060075461864471, 0.030552517622709274, -0.0934130847454071, 0.03580627590417862, 0.06884689629077911, -0.023429861292243004, -0.2501198947429657, -0.04468001052737236, -0.04278911277651787, -0.09598927944898605, 0.10400696098804474, 0.0502825565636158, -0.04907103255391121, 0.04504842311143875, -0.00390124530531466, 0.027359874919056892, -0.023355862125754356, 0.0648031160235405, 0.07551190257072449, 0.05339067056775093, 0.09936446696519852, -0.029688572511076927, -0.02128875069320202, 0.07469788193702698, 0.0016285483725368977, 0.2729965150356293, -0.025124449282884598, 0.08217522501945496, 0.029772909358143806, 0.09764909744262695, -0.015689224004745483, 0.054648540914058685, 0.01913999207317829, -0.006189570762217045, -0.012593789957463741, -0.05925280600786209, -0.03667555749416351, 0.029177982360124588, -0.008866158314049244, 0.0005986399482935667, -0.08624964952468872, 0.05615432560443878, 0.02696870267391205, 0.23044781386852264, 0.015929311513900757, -0.2859264314174652, -0.05387839302420616, -0.00032510937307961285, -0.027619360014796257, -0.06736818701028824, -0.0004877043829765171, 0.11634249240159988, -0.11007877439260483, 0.07052038609981537, -0.06447853147983551, 0.08418519049882889, -0.048696525394916534, -0.009468977339565754, 0.05537819862365723, 0.16350752115249634, -0.01264458242803812, 0.06557376682758331, -0.20758561789989471, 0.21625232696533203, 0.02283305674791336, 0.12419125437736511, -0.07833241671323776, 0.042894769459962845, 0.012037585489451885, 0.00723566859960556, 0.07642718404531479, -0.0003850124485325068, -0.12651176750659943, -0.12972354888916016, -0.08250851184129715, 0.06965956091880798, 0.15671998262405396, -0.017590558156371117, 0.08193589746952057, -0.04237138852477074, 0.014004180207848549, 0.04981786385178566, -0.097343809902668, -0.16306860744953156, -0.1617289036512375, 0.025300439447164536, 0.015017199330031872, -0.03631508722901344, -0.051722537726163864, -0.09163691103458405, -0.021953992545604706, 0.16045571863651276, -0.009273878298699856, -0.03911059722304344, -0.15946689248085022, 0.08041573315858841, 0.16748172044754028, -0.055151067674160004, 0.02771769091486931, 0.011783869005739689, 0.11250342428684235, 0.03423618897795677, -0.09000343829393387, 0.057259391993284225, -0.07457638531923294, -0.15445342659950256, -0.060682255774736404, 0.09994236379861832, 0.05986848473548889, 0.0398382805287838, -0.017434990033507347, 0.04452921450138092, -0.014743051491677761, -0.09640920907258987, 0.028687624260783195, 0.09789053350687027, 0.071155846118927, 0.06710581481456757, -0.09414729475975037, 0.040657054632902145, -0.01066980604082346, -0.03466583043336868, 0.13535581529140472, 0.20062635838985443, -0.07307947427034378, 0.1007634773850441, 0.08051475137472153, -0.08678901940584183, -0.18313388526439667, 0.07723983377218246, 0.10982122272253036, 0.023353781551122665, 0.04618017002940178, -0.2444296032190323, 0.1284516155719757, 0.1235194206237793, -0.004735278896987438, 0.04910382628440857, -0.2962914705276489, -0.12382427603006363, 0.05130846053361893, 0.10463564842939377, 0.027785349637269974, -0.11879131197929382, -0.024223580956459045, -0.04195600003004074, -0.11118153482675552, 0.1489134579896927, -0.07719110697507858, 0.11739509552717209, 0.004122366663068533, 0.07835808396339417, 0.02186555042862892, -0.04507853835821152, 0.1281244307756424, 0.022837115451693535, 0.06832530349493027, -0.03176300972700119, 0.013868384063243866, 0.0041277650743722916, -0.05668235942721367, 0.0280451700091362, -0.09168116748332977, 0.0677982047200203, -0.10789845883846283, -0.01810794696211815, -0.048611145466566086, 0.05286063998937607, -0.04058549925684929, -0.05919503793120384, -0.04309127852320671, 0.04512215405702591, 0.07602186501026154, -0.03339163959026337, 0.06979021430015564, 0.0064103612676262856, 0.07118088752031326, 0.03070315159857273, 0.09834277629852295, -0.052532121539115906, -0.047829046845436096, 0.001995707396417856, -0.011556003242731094, 0.04979119449853897, -0.1282782107591629, 0.022182002663612366, 0.13007891178131104, 0.042189646512269974, 0.1326419711112976, 0.0548853725194931, -0.04642464593052864, 0.000224738716497086, 0.05458470433950424, -0.1174076572060585, -0.11846155673265457, 0.008663350716233253, -0.039474520832300186, -0.1266569048166275, 0.013953850604593754, 0.10537197440862656, -0.02862364798784256, -0.008908191695809364, -0.018367232754826546, 0.03584634140133858, 0.00475693866610527, 0.19043882191181183, 0.012109681032598019, 0.050455063581466675, -0.10563483834266663, 0.14386236667633057, 0.055652230978012085, -0.1041230857372284, 0.07363030314445496, 0.08936110138893127, -0.08635934442281723, 0.0002597303537186235, 0.06285940855741501, 0.14040826261043549, -0.03198086470365524, -0.048121221363544464, -0.0911392867565155, -0.09235911816358566, 0.059958625584840775, 0.09767132997512817, 0.02685694210231304, 0.005462441593408585, -0.04935876652598381, 0.03410130366683006, -0.1528376042842865, 0.06469488143920898, 0.0538882277905941, 0.06383003294467926, -0.10762068629264832, 0.10195342451334, 0.019678102806210518, 0.004638612270355225, -0.012200725264847279, 0.013673083856701851, -0.09949826449155807, -0.026570547372102737, -0.06929156929254532, -0.011723389849066734, -0.03303110599517822, -0.0013672139029949903, -0.0020270314998924732, -0.03729248046875, -0.05925155058503151, 0.03271092474460602, -0.07692810148000717, -0.059513386338949203, -0.009115982800722122, 0.042971786111593246, -0.13201060891151428, 0.010381367988884449, 0.017324917018413544, -0.10111100226640701, 0.07913587987422943, 0.06269128620624542, 0.02122356928884983, 0.0360582210123539, -0.15330202877521515, -0.04023728519678116, 0.028421157971024513, 0.018358752131462097, 0.06032971665263176, -0.08942531049251556, -0.007819751277565956, -0.005652437452226877, 0.04794618487358093, 0.009375927038490772, 0.057465218007564545, -0.11569656431674957, -0.0026330447290092707, -0.0792798101902008, -0.05318831279873848, -0.06370829790830612, 0.04510390758514404, 0.10680972039699554, 0.042908135801553726, 0.16045919060707092, -0.06244076043367386, 0.027356285601854324, -0.18475212156772614, -0.02042044699192047, -0.007134201005101204, -0.038040656596422195, -0.09167850017547607, -0.023693028837442398, 0.09068803489208221, -0.04887191206216812, 0.13985443115234375, 0.0056189922615885735, 0.05383681878447533, 0.033586129546165466, -0.02232315205037594, -0.04365126043558121, 0.007992894388735294, 0.205439493060112, 0.08427152782678604, -0.01701914519071579, 0.08888769149780273, 0.004146704450249672, 0.06333114951848984, 0.05933554098010063, 0.22004687786102295, 0.15713392198085785, 0.009619554504752159, 0.08687067031860352, 0.05285925790667534, -0.12851791083812714, -0.10328669846057892, 0.16639070212841034, -0.05445007234811783, 0.11786209046840668, -0.043076131492853165, 0.18683789670467377, 0.06960698217153549, -0.16706393659114838, 0.02670513093471527, -0.03975346311926842, -0.09953851997852325, -0.13193966448307037, -0.0027049125637859106, -0.08969219774007797, -0.11514025926589966, 0.02857867069542408, -0.13019219040870667, 0.051600802689790726, 0.07488037645816803, 0.019214637577533722, 0.03439921513199806, 0.10300813615322113, -0.025800317525863647, -0.0007605129503645003, 0.047301508486270905, 0.02962210401892662, -0.003324556164443493, -0.04651289060711861, -0.0921434536576271, 0.03370389714837074, 0.01110687106847763, 0.09482290595769882, -0.05037442594766617, -0.01059427298605442, 0.054555732756853104, -0.013547680340707302, -0.06211091950535774, 0.019097672775387764, 0.009767584502696991, 0.04115238040685654, 0.04086829721927643, 0.04596560448408127, -0.009496110491454601, -0.044488921761512756, 0.27219995856285095, -0.06421928852796555, -0.06603933870792389, -0.13492253422737122, 0.1940087080001831, 0.04100622981786728, -0.010490350425243378, 0.06853989511728287, -0.10560283809900284, -0.041589464992284775, 0.16170720756053925, 0.1634044051170349, -0.0725482627749443, -0.03292755410075188, -0.0029975229408591986, -0.010107534937560558, -0.04990864545106888, 0.13934707641601562, 0.09357026219367981, 0.06617047637701035, -0.04806935787200928, -0.018128301948308945, -0.01986832171678543, -0.043899212032556534, -0.05302096903324127, 0.05767527595162392, 0.024204755201935768, -0.01768294908106327, -0.02401925064623356, 0.07225306332111359, -0.013864408247172832, -0.18351823091506958, 0.03193776682019234, -0.1719445288181305, -0.1781451553106308, -0.020380983129143715, 0.09507531672716141, -0.02659040316939354, 0.04929323121905327, -0.00297738891094923, -0.018864192068576813, 0.09232476353645325, -0.004631711635738611, -0.0605299137532711, -0.1021454930305481, 0.07601436227560043, -0.06451287865638733, 0.21415898203849792, -0.008751694113016129, 0.06076151877641678, 0.10287729650735855, 0.052888091653585434, -0.12861499190330505, 0.026416471228003502, 0.05824212357401848, -0.09783870726823807, 0.04145854339003563, 0.14231479167938232, -0.06563084572553635, 0.07761086523532867, 0.048650406301021576, -0.09372500330209732, 0.0030410769395530224, -0.07436812669038773, -0.03684881702065468, -0.05525444447994232, -0.015084439888596535, -0.06789664179086685, 0.16865085065364838, 0.20665593445301056, -0.015402726829051971, 0.012876832857728004, -0.07557801902294159, 0.042174987494945526, 0.03455112874507904, 0.09723647683858871, -0.03288199380040169, -0.23031504452228546, 0.01564560830593109, 0.03818404674530029, 0.014788844622671604, -0.20606949925422668, -0.07323551923036575, 0.027115561068058014, -0.05025370791554451, -0.06273557990789413, 0.10266801714897156, 0.06519389897584915, 0.025007877498865128, -0.04419024661183357, -0.1097111776471138, -0.02202746272087097, 0.13874448835849762, -0.14966103434562683, -0.04047732427716255 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta_roberta_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "roberta_roberta_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/roberta_roberta_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# roberta_roberta_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
[ "# roberta_roberta_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# roberta_roberta_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ 62, 40, 6, 12, 8, 3, 118, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# roberta_roberta_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ -0.08111652731895447, 0.13502509891986847, -0.003910874016582966, 0.06949828565120697, 0.14037403464317322, 0.03863140195608139, 0.10616826266050339, 0.1152031198143959, -0.11491961032152176, 0.05421384051442146, 0.07564882934093475, 0.07000955194234848, 0.0438128300011158, 0.15776333212852478, -0.030419858172535896, -0.22096072137355804, -0.0002738511248026043, -0.011529305018484592, -0.08427827060222626, 0.11876288056373596, 0.08497526496648788, -0.10052355378866196, 0.05886903405189514, -0.004022842273116112, -0.13675916194915771, 0.02218209207057953, -0.04216821864247322, -0.051906805485486984, 0.09625883400440216, -0.0022812006063759327, 0.09339755028486252, 0.03227158263325691, 0.14547118544578552, -0.2014598250389099, 0.005891341716051102, 0.09823357313871384, 0.039517004042863846, 0.09511525928974152, 0.05680086091160774, -0.02038070559501648, 0.13996988534927368, -0.14981083571910858, 0.10214269161224365, 0.019049732014536858, -0.07810165733098984, -0.12725642323493958, -0.1026802659034729, 0.053845565766096115, 0.0844467505812645, 0.120578832924366, 0.005850763060152531, 0.1354817897081375, -0.11186349391937256, 0.08276455104351044, 0.17911690473556519, -0.2048257291316986, -0.049327488988637924, 0.06128178536891937, 0.023980148136615753, 0.05766971781849861, -0.09345655888319016, -0.017841454595327377, 0.021086890250444412, 0.021793020889163017, 0.09829448163509369, 0.0013069846900179982, -0.09781330078840256, 0.00006225322431419045, -0.12374068796634674, -0.019055884331464767, 0.08021371066570282, 0.023556062951683998, -0.009169606491923332, -0.09175100922584534, -0.056515660136938095, -0.1400512158870697, -0.022936556488275528, -0.007094166707247496, 0.03717825561761856, -0.056958895176649094, -0.07563553750514984, -0.04179872199892998, -0.06453460454940796, -0.04799901694059372, -0.024889137595891953, 0.12887324392795563, 0.046800997108221054, 0.009882288984954357, -0.04427291080355644, 0.11487879604101181, 0.07337695360183716, -0.12259776890277863, 0.01096566766500473, 0.011937382631003857, -0.09692306816577911, -0.04764564335346222, -0.025740617886185646, -0.02119414135813713, -0.005378857720643282, 0.14183059334754944, -0.07130506634712219, 0.08049659430980682, 0.024239590391516685, -0.014980954118072987, -0.017644846811890602, 0.14703024923801422, -0.048129163682460785, -0.046376489102840424, -0.008443755097687244, 0.10620619356632233, 0.008089783601462841, -0.01136521715670824, -0.06699658185243607, 0.0072926683351397514, 0.06262866407632828, 0.0639452189207077, -0.039873965084552765, 0.04680204391479492, -0.01508901547640562, -0.015088105574250221, 0.04338531196117401, -0.13642191886901855, 0.037628788501024246, 0.017244255170226097, -0.10156180709600449, -0.014366962015628815, 0.017918383702635765, -0.00478913402184844, -0.045214246958494186, 0.12510839104652405, -0.07639573514461517, 0.006566964089870453, -0.0821308121085167, -0.07888142764568329, 0.009789477102458477, -0.09810982644557953, -0.034859590232372284, -0.036642417311668396, -0.17672693729400635, -0.060545507818460464, 0.02783540077507496, -0.061928801238536835, -0.036596812307834625, -0.053994517773389816, -0.07683632522821426, 0.02243700437247753, -0.006975103169679642, 0.18445535004138947, -0.06798209249973297, 0.06955590844154358, -0.019562603905797005, 0.02798369526863098, 0.06594757735729218, 0.04589637741446495, -0.07287628948688507, -0.010335089638829231, -0.10135944932699203, 0.08786599338054657, -0.08099188655614853, 0.011188448406755924, -0.11018364131450653, -0.09362151473760605, 0.02100781537592411, -0.008547577075660229, 0.05433927848935127, 0.12565775215625763, -0.18991202116012573, -0.03577248379588127, 0.11848165094852448, -0.05380861833691597, -0.006994294933974743, 0.05076717585325241, -0.05685477703809738, -0.0035720281302928925, 0.06238443776965141, 0.1220242902636528, 0.06823934614658356, -0.11713501065969467, -0.0037190131843090057, -0.0020233220420777798, 0.030535127967596054, 0.008169758133590221, 0.01856476627290249, 0.006308402866125107, 0.062320902943611145, 0.007000882178544998, -0.03560139238834381, 0.014771017245948315, -0.09690316766500473, -0.06526549160480499, -0.03478359058499336, -0.07550253719091415, 0.009467991068959236, 0.03206581622362137, 0.03392494097352028, -0.05936473235487938, -0.1175139918923378, 0.07567503303289413, 0.128676638007164, -0.055855583399534225, 0.014438631944358349, -0.07550030946731567, -0.04019228368997574, 0.012662711553275585, -0.018947381526231766, -0.19975906610488892, -0.11025889962911606, 0.0180598646402359, -0.05680741369724274, 0.04503960907459259, -0.011639255098998547, 0.0652574896812439, 0.04348974674940109, -0.03794136643409729, -0.01714145392179489, -0.05996111407876015, 0.0017921014223247766, -0.08678042888641357, -0.21295066177845, -0.030305111780762672, -0.01876487210392952, 0.16297246515750885, -0.218558669090271, 0.012140180915594101, -0.02230932004749775, 0.14703750610351562, 0.02707231417298317, -0.05665484070777893, 0.016224386170506477, 0.05152866989374161, 0.006102025508880615, -0.0956352949142456, 0.030176285654306412, -0.01040698029100895, -0.07524880021810532, -0.010428059846162796, -0.14146991074085236, -0.023198992013931274, 0.0803331583738327, 0.09192106872797012, -0.1068134531378746, 0.026156244799494743, -0.07443752884864807, -0.041072238236665726, -0.08634673058986664, 0.04045098274946213, 0.2019098699092865, 0.03491226211190224, 0.13251733779907227, -0.04588942602276802, -0.08516594022512436, -0.009075822308659554, 0.01999027654528618, 0.015936965122818947, 0.11143715679645538, 0.08136831969022751, -0.05566449463367462, 0.07308534532785416, 0.013184533454477787, -0.043000172823667526, 0.1328134983778, -0.044834285974502563, -0.08551783859729767, 0.0009717674693092704, -0.03061283938586712, -0.022371452301740646, 0.09735188633203506, -0.02649461291730404, 0.00572229316458106, 0.03320265933871269, 0.024044889956712723, 0.0329449288547039, -0.18195585906505585, -0.0009482029709033668, 0.013233584351837635, -0.056180451065301895, -0.03640580549836159, -0.015520550310611725, 0.0544341579079628, 0.10241883993148804, 0.004297931678593159, -0.03288105130195618, 0.007464804220944643, -0.005949416663497686, -0.06781370937824249, 0.1763957142829895, -0.11147643625736237, -0.16417737305164337, -0.08016534149646759, 0.01536510419100523, -0.024689434096217155, -0.052205026149749756, 0.004359261132776737, -0.11707401275634766, -0.07273957133293152, -0.11218053102493286, -0.012387686409056187, 0.007409687153995037, -0.002833086997270584, 0.04086765646934509, 0.019475489854812622, 0.04932611435651779, -0.13798265159130096, 0.016637522727251053, -0.05473756045103073, -0.07873135805130005, 0.021493617445230484, 0.07797432690858841, 0.06182488799095154, 0.14682097733020782, -0.014928515069186687, 0.029302826151251793, -0.017045622691512108, 0.19359512627124786, -0.09432525187730789, 0.013021022081375122, 0.1030004620552063, 0.01978609338402748, 0.042137935757637024, 0.12277855724096298, 0.03942249342799187, -0.0755474716424942, 0.030830560252070427, 0.0970337837934494, -0.025676459074020386, -0.26007914543151855, -0.06316322833299637, -0.01861269772052765, -0.08620913326740265, 0.08180374652147293, 0.057834405452013016, -0.011996099725365639, 0.006424378138035536, -0.010263586416840553, -0.01693420112133026, -0.0030224148649722338, 0.05295253545045853, 0.07961849123239517, 0.05734461173415184, 0.09430722892284393, -0.03831564635038376, -0.02820059470832348, 0.06368395686149597, -0.005868020001798868, 0.23096659779548645, -0.06675843894481659, 0.07089507579803467, 0.022569816559553146, 0.086217500269413, -0.011266044341027737, 0.02501630410552025, 0.02633858658373356, -0.013448154553771019, 0.011427310295403004, -0.045976292341947556, 0.0008616389823146164, -0.0002786587574519217, -0.015601416118443012, -0.0035510531160980463, -0.05958838015794754, 0.04110454022884369, 0.014756634831428528, 0.27500128746032715, 0.033544719219207764, -0.2870272397994995, -0.053007371723651886, -0.02303057722747326, -0.034437041729688644, -0.042568985372781754, -0.009805022738873959, 0.08729251474142075, -0.11466723680496216, 0.06389641016721725, -0.04382767900824547, 0.08362310379743576, -0.07627679407596588, -0.0035400334745645523, 0.06381712853908539, 0.14925143122673035, -0.017811184749007225, 0.060678623616695404, -0.21177631616592407, 0.22443312406539917, 0.00592814153060317, 0.12357211858034134, -0.06154519319534302, 0.018854895606637, 0.018126526847481728, 0.03819498419761658, 0.06126532703638077, -0.007346613332629204, -0.030703788623213768, -0.15418371558189392, -0.11525250971317291, 0.026326164603233337, 0.11508464813232422, 0.0010356673737987876, 0.0839175283908844, -0.02805805765092373, -0.01751401647925377, 0.03842596337199211, -0.1641320437192917, -0.15152817964553833, -0.12486698478460312, 0.03022219054400921, 0.028833836317062378, -0.04564567655324936, -0.05695609748363495, -0.1154336929321289, -0.01953401230275631, 0.17371566593647003, 0.026928162202239037, -0.05940547585487366, -0.1462540179491043, 0.06677138060331345, 0.15283823013305664, -0.04580932855606079, 0.009701916016638279, 0.031702443957328796, 0.11238211393356323, 0.04849906265735626, -0.08543648570775986, 0.0366286039352417, -0.06118723005056381, -0.18790753185749054, -0.05417255312204361, 0.12725965678691864, 0.08923226594924927, 0.050085317343473434, -0.024312596768140793, 0.030653303489089012, -0.0025363669265061617, -0.07853776961565018, 0.006540127098560333, 0.07986976951360703, 0.0703912302851677, 0.06173519790172577, -0.06311336904764175, -0.003706437535583973, -0.04503060504794121, -0.025167254731059074, 0.10399968177080154, 0.1916821151971817, -0.07961055636405945, 0.10967834293842316, 0.061961278319358826, -0.0647507756948471, -0.17617200314998627, 0.05265549197793007, 0.12391509115695953, 0.00933986995369196, 0.031848277896642685, -0.20430263876914978, 0.11962920427322388, 0.11788742989301682, -0.011639069765806198, 0.04798618704080582, -0.34696346521377563, -0.12411915510892868, 0.05770426243543625, 0.08325227349996567, 0.020814992487430573, -0.10562187433242798, -0.019959570840001106, -0.015137789770960808, -0.13509255647659302, 0.13725866377353668, -0.06297162920236588, 0.11435330659151077, -0.009754933416843414, 0.10358688980340958, 0.022013643756508827, -0.04305771738290787, 0.1031682938337326, 0.05843938887119293, 0.07235973328351974, -0.04641496390104294, 0.019572336226701736, 0.0407947413623333, -0.06268996745347977, 0.03459566459059715, -0.04752090945839882, 0.07140462100505829, -0.11518732458353043, -0.013878284022212029, -0.08506715297698975, 0.04332845285534859, -0.04128525033593178, -0.04277094453573227, -0.04339263215661049, 0.05397909879684448, 0.07357032597064972, -0.046627387404441833, 0.07036546617746353, 0.0018686003750190139, 0.0835462361574173, 0.06159831956028938, 0.08438918739557266, -0.04306608438491821, -0.08919881284236908, -0.008973811753094196, -0.009121058508753777, 0.030424386262893677, -0.12272141128778458, 0.022938715294003487, 0.1308310627937317, 0.04890736937522888, 0.11988747864961624, 0.03378443419933319, -0.03984139859676361, -0.015707895159721375, 0.038525182753801346, -0.11659369617700577, -0.08896790444850922, 0.03289111703634262, -0.06666748225688934, -0.10492413491010666, 0.013613667339086533, 0.10462965071201324, -0.034142643213272095, -0.021321216598153114, 0.0029425721149891615, 0.030421707779169083, 0.015017312951385975, 0.21684592962265015, 0.029750436544418335, 0.06463317573070526, -0.11676184087991714, 0.12856978178024292, 0.06341592967510223, -0.07907187938690186, 0.04594944044947624, 0.13539637625217438, -0.1012483686208725, -0.01303454115986824, 0.09757645428180695, 0.1384163349866867, -0.038327138870954514, -0.027378063648939133, -0.09771161526441574, -0.09334137290716171, 0.07975511252880096, 0.13207311928272247, 0.030626792460680008, 0.0016679763793945312, -0.05250655487179756, 0.013256015256047249, -0.1521102786064148, 0.06777632236480713, 0.0841919481754303, 0.05343497171998024, -0.08232088387012482, 0.1302136331796646, 0.02720477804541588, 0.031472790986299515, -0.01777525246143341, 0.012873890809714794, -0.07321316003799438, -0.009206348098814487, -0.08120190352201462, -0.013421202078461647, -0.008489429019391537, -0.0016425480134785175, -0.022571532055735588, -0.052527859807014465, -0.04284016415476799, 0.045717038214206696, -0.07675134390592575, -0.05875604227185249, -0.0032794629223644733, 0.03979533910751343, -0.1366533637046814, 0.0020425317343324423, 0.0054398188367486, -0.09166456013917923, 0.0714287981390953, 0.04721227288246155, -0.00427798880264163, 0.0325496643781662, -0.15704987943172455, -0.04372134804725647, 0.029711108654737473, 0.012727262452244759, 0.07790299504995346, -0.07847821712493896, -0.00021524183102883399, -0.02120085619390011, 0.04190433770418167, 0.014804619364440441, 0.0495804101228714, -0.11488575488328934, 0.010123956017196178, -0.07381915301084518, -0.03310724347829819, -0.07039649784564972, 0.042319249361753464, 0.1060481071472168, 0.04478493705391884, 0.1763048619031906, -0.08088657259941101, 0.03744947537779808, -0.18233387172222137, -0.030538659542798996, -0.006076015532016754, -0.04343516752123833, -0.07402870804071426, -0.013314413838088512, 0.10870972275733948, -0.05200359597802162, 0.12433113902807236, 0.012726617977023125, 0.06706410646438599, 0.042196955531835556, -0.025109155103564262, -0.06375615298748016, 0.012713083066046238, 0.140381321310997, 0.06253359466791153, -0.018374042585492134, 0.11606087535619736, -0.012407327070832253, 0.048397649079561234, 0.04202011972665787, 0.22265610098838806, 0.13464562594890594, 0.03521754592657089, 0.06730546057224274, 0.047094766050577164, -0.14080023765563965, -0.10849335044622421, 0.11338698118925095, -0.09022855758666992, 0.12030712515115738, -0.06407644599676132, 0.20509304106235504, 0.06354007124900818, -0.13940273225307465, 0.05263267830014229, -0.030122457072138786, -0.09963593631982803, -0.10560356080532074, -0.006256684195250273, -0.0803617611527443, -0.10861707478761673, 0.03900912404060364, -0.1096944659948349, 0.0644855871796608, 0.0986337661743164, 0.02969737909734249, 0.02811264432966709, 0.10519406199455261, -0.00011908977467101067, -0.0033962891902774572, 0.05422317609190941, 0.02291373908519745, -0.006015505641698837, -0.0625951737165451, -0.06228799372911453, 0.02778146043419838, 0.024392269551753998, 0.08391272276639938, -0.03424936532974243, -0.00864073820412159, 0.04789908975362778, -0.023515040054917336, -0.07626352459192276, 0.04009772092103958, 0.007934189401566982, 0.06691183894872665, 0.056678567081689835, 0.05545781925320625, 0.005330403335392475, -0.047101981937885284, 0.29454949498176575, -0.061694446951150894, -0.11447452753782272, -0.12434607744216919, 0.23757004737854004, 0.057765327394008636, -0.026035195216536522, 0.0604732371866703, -0.10356482118368149, -0.04433297738432884, 0.15712283551692963, 0.17416979372501373, -0.054571930319070816, -0.026691677048802376, -0.03202073648571968, -0.012960435822606087, -0.028282076120376587, 0.12446622550487518, 0.1143776923418045, 0.052006497979164124, -0.04919760301709175, -0.024272913113236427, -0.019418496638536453, -0.04046645760536194, -0.07007301598787308, 0.07202598452568054, 0.020590534433722496, -0.005499145481735468, -0.035142917186021805, 0.07126759737730026, -0.014898622408509254, -0.21098951995372772, 0.03973717615008354, -0.1660938709974289, -0.18607185781002045, -0.008858168497681618, 0.09379781037569046, -0.021586086601018906, 0.047220028936862946, 0.01629958674311638, -0.016059977933764458, 0.09929732233285904, -0.029422370716929436, -0.03775728866457939, -0.10529567301273346, 0.06422360241413116, -0.1289294809103012, 0.22625763714313507, -0.011120143346488476, 0.06397224217653275, 0.09556955099105835, 0.020884616300463676, -0.12811550498008728, 0.06737964600324631, 0.04410437121987343, -0.09633247554302216, 0.03498329594731331, 0.1637641042470932, -0.04153206944465637, 0.047182779759168625, 0.039424996823072433, -0.11532149463891983, -0.0010438188910484314, -0.09230833500623703, -0.030374081805348396, -0.06062634661793709, -0.017300672829151154, -0.027894284576177597, 0.14977137744426727, 0.21635878086090088, -0.009171972051262856, 0.03050398826599121, -0.08351695537567139, -0.006214710418134928, 0.026010602712631226, 0.0785079225897789, -0.029073115438222885, -0.21976760029792786, 0.03246847540140152, 0.013044595718383789, 0.036313630640506744, -0.18713438510894775, -0.06949179619550705, 0.002247417811304331, -0.07088150084018707, -0.05087396875023842, 0.11307258903980255, 0.047705378383398056, 0.042886991053819656, -0.02392316795885563, -0.08030718564987183, -0.010341732762753963, 0.14629141986370087, -0.19319826364517212, -0.03680497407913208 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # robertagpt2_cnn This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.13.0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "robertagpt2_cnn", "results": []}]}
text2text-generation
Ayham/robertagpt2_cnn
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
# robertagpt2_cnn This model is a fine-tuned version of [](URL on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.13.0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# robertagpt2_cnn\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.13.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n", "# robertagpt2_cnn\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.13.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 53, 31, 6, 12, 8, 3, 118, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n# robertagpt2_cnn\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.13.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.0649973601102829, 0.13222366571426392, -0.0033052372746169567, 0.05813542753458023, 0.13808812201023102, 0.03283016011118889, 0.1055135726928711, 0.12075625360012054, -0.09052843600511551, 0.09000866860151291, 0.08742600679397583, 0.07161498069763184, 0.05433835834264755, 0.11556697636842728, -0.02641645073890686, -0.2510875165462494, 0.03222031146287918, -0.0029344838112592697, -0.08063769340515137, 0.09213852882385254, 0.09956159442663193, -0.09853878617286682, 0.0682985782623291, 0.00884173158556223, -0.14953827857971191, 0.01888139918446541, -0.05688343569636345, -0.06877546012401581, 0.07947345077991486, 0.02015777863562107, 0.0888170674443245, 0.022817449644207954, 0.09267544001340866, -0.2382226139307022, 0.006849030498415232, 0.06526414304971695, 0.02384583279490471, 0.08401445299386978, 0.07036596536636353, -0.013952518813312054, 0.12369251251220703, -0.14502239227294922, 0.11004117131233215, 0.027679957449436188, -0.06201431155204773, -0.1407763808965683, -0.08289086073637009, 0.07888276875019073, 0.07607334852218628, 0.08392904698848724, 0.00879550538957119, 0.1249946877360344, -0.10811823606491089, 0.07632911950349808, 0.1917446404695511, -0.24970701336860657, -0.05677700787782669, 0.022081343457102776, 0.07071293890476227, 0.06735993921756744, -0.09728457778692245, -0.00805610604584217, 0.039847467094659805, 0.034659333527088165, 0.08904606103897095, 0.0025157069321721792, -0.07240474224090576, 0.002497141482308507, -0.13169533014297485, -0.034241825342178345, 0.14102131128311157, 0.051794763654470444, -0.035450659692287445, -0.1254749298095703, -0.03743869066238403, -0.10992862284183502, -0.018420642241835594, -0.036497849971055984, 0.03450552746653557, -0.038155991584062576, -0.0917198657989502, -0.060544952750205994, -0.06405261904001236, -0.05800742283463478, -0.004804433789104223, 0.16189968585968018, 0.05532724782824516, -0.0027066045440733433, -0.03664933517575264, 0.09639177471399307, 0.057640835642814636, -0.11099312454462051, -0.019942909479141235, -0.008030736818909645, -0.07188097387552261, -0.050429608672857285, -0.0577978752553463, -0.05284682661294937, 0.03059394285082817, 0.18404777348041534, -0.058226391673088074, 0.09920470416545868, 0.01028440147638321, 0.007682365365326405, -0.00029148353496566415, 0.12139453738927841, -0.040901970118284225, -0.01541177649050951, -0.01404380053281784, 0.09317059069871902, 0.014671210199594498, -0.023973142728209496, -0.08891483396291733, -0.028814077377319336, 0.09898808598518372, 0.06641692668199539, -0.025756223127245903, 0.03171764686703682, -0.04844953119754791, -0.031299956142902374, 0.013912606053054333, -0.11796839535236359, 0.03954056277871132, 0.007644038647413254, -0.0775342509150505, 0.015944983810186386, 0.026029368862509727, 0.005642310716211796, -0.04132803529500961, 0.08691153675317764, -0.07997661083936691, -0.01567448303103447, -0.0713437870144844, -0.06329745799303055, 0.036146219819784164, -0.06552141904830933, -0.004871795419603586, -0.0845651924610138, -0.14986340701580048, -0.03934277594089508, 0.04009062796831131, -0.050434958189725876, -0.03906421735882759, -0.052774928510189056, -0.046314552426338196, 0.042747270315885544, -0.018015895038843155, 0.13168518245220184, -0.05207604542374611, 0.06967882812023163, -0.015314093790948391, 0.03590553626418114, 0.03448839485645294, 0.04202340543270111, -0.06697045266628265, 0.025466080754995346, -0.11599671840667725, 0.09219695627689362, -0.09633810818195343, 0.015901677310466766, -0.1193215474486351, -0.09402008354663849, -0.013543374836444855, -0.002338313264772296, 0.0747009813785553, 0.12713126838207245, -0.10849487036466599, -0.0489993654191494, 0.151204451918602, -0.060564398765563965, -0.06864472478628159, 0.09449892491102219, -0.04560043662786484, -0.01800260879099369, 0.0528140552341938, 0.13662965595722198, 0.09149648994207382, -0.11112789064645767, -0.031215261667966843, 0.026616619899868965, 0.05488136410713196, 0.01828271895647049, 0.06857912242412567, -0.01697365939617157, -0.007311782333999872, 0.02099461853504181, -0.029032927006483078, 0.027662916108965874, -0.08777663856744766, -0.07818751782178879, -0.046821318566799164, -0.07511181384325027, 0.050301145762205124, 0.0014205619227141142, 0.04030458256602287, -0.07533354312181473, -0.12402530759572983, 0.1067802906036377, 0.1378578394651413, -0.07377076894044876, 0.024685578420758247, -0.07162847369909286, 0.018814926967024803, -0.028344929218292236, -0.014933492057025433, -0.20251944661140442, -0.10670872032642365, 0.03904392570257187, -0.09202471375465393, 0.0496414490044117, -0.023081906139850616, 0.06528709828853607, 0.05539059266448021, -0.03224334120750427, -0.04264642670750618, -0.09402837604284286, -0.002063618041574955, -0.0928342193365097, -0.19285985827445984, -0.04924742504954338, -0.03687519580125809, 0.14955642819404602, -0.19400420784950256, 0.009423201903700829, 0.008227088488638401, 0.17293527722358704, 0.027650360018014908, -0.04626632481813431, 0.009788816794753075, 0.054604724049568176, -0.017871705815196037, -0.08043611794710159, 0.03530475124716759, 0.005843666847795248, -0.1329224556684494, -0.03159273788332939, -0.12487819790840149, 0.026068810373544693, 0.09545715153217316, 0.030432987958192825, -0.0874306708574295, -0.03961320221424103, -0.0673157200217247, -0.047349318861961365, -0.04767214506864548, -0.0013044393854215741, 0.19221866130828857, 0.01696299947798252, 0.13102035224437714, -0.057239364832639694, -0.06740991771221161, 0.0034078070893883705, 0.008734527975320816, -0.03793301805853844, 0.09569232165813446, 0.07524625211954117, -0.07758302986621857, 0.08349409699440002, 0.08699601143598557, -0.05288566276431084, 0.15303239226341248, -0.047509025782346725, -0.09499914199113846, -0.015535901300609112, 0.03215767443180084, -0.008436186239123344, 0.06328309327363968, -0.10663598030805588, -0.0031234202906489372, 0.02947407402098179, 0.02788025140762329, 0.05613034591078758, -0.15841589868068695, 0.014159084297716618, 0.02353653311729431, -0.03527205437421799, 0.01393734011799097, -0.007534128613770008, 0.024248482659459114, 0.08069317787885666, 0.04949605464935303, -0.006566836964339018, 0.03255755826830864, -0.0062959156930446625, -0.07610002160072327, 0.1873161345720291, -0.13778060674667358, -0.16931580007076263, -0.1331237256526947, 0.03579702600836754, -0.06733144074678421, -0.03280067443847656, 0.02607494220137596, -0.10058057308197021, -0.06077120080590248, -0.07527109980583191, -0.008067520335316658, -0.08319524675607681, 0.009033170528709888, 0.06203581392765045, 0.01784222573041916, 0.056240782141685486, -0.1385006159543991, 0.01176324300467968, -0.013075727969408035, -0.07073129713535309, 0.0031396776903420687, 0.05044860020279884, 0.09870531409978867, 0.12006635218858719, 0.0005658300360664725, 0.014768059365451336, -0.03348967432975769, 0.20458300411701202, -0.0687103420495987, -0.03176608681678772, 0.10532404482364655, -0.00040996598545461893, 0.06261394917964935, 0.07942413538694382, 0.037188347429037094, -0.08058609068393707, 0.038689035922288895, 0.05805753916501999, -0.007147603668272495, -0.2615112066268921, -0.04661088436841965, -0.037279073148965836, -0.06673956662416458, 0.10373307764530182, 0.06207649037241936, 0.003212034236639738, 0.048608917742967606, -0.0260305292904377, 0.05443647876381874, -0.01794143207371235, 0.09180842339992523, 0.10128378123044968, 0.04430849850177765, 0.08668557554483414, -0.03333692252635956, -0.03753291815519333, 0.07320673018693924, 0.007908889092504978, 0.2421024888753891, -0.014379523694515228, 0.13923080265522003, 0.03409755229949951, 0.1195906475186348, -0.012526732869446278, 0.026429427787661552, 0.022962436079978943, 0.0154677489772439, -0.000799414177890867, -0.06219395250082016, -0.035273391753435135, 0.026766033843159676, 0.019523823633790016, 0.024686520919203758, -0.10829009860754013, 0.023151330649852753, 0.025387678295373917, 0.26153677701950073, 0.008470551110804081, -0.28757184743881226, -0.06993088126182556, 0.018598301336169243, -0.03519441932439804, -0.061516501009464264, 0.015493305400013924, 0.09031432867050171, -0.13335345685482025, 0.06343026459217072, -0.050148822367191315, 0.09211789816617966, -0.06470289826393127, -0.009142049588263035, 0.020480012521147728, 0.15255191922187805, -0.001224344247020781, 0.08321842551231384, -0.21518753468990326, 0.18309526145458221, 0.015388930216431618, 0.10736019909381866, -0.06479739397764206, 0.04491173475980759, 0.021826529875397682, 0.032280102372169495, 0.08580341935157776, -0.00835180189460516, -0.09635044634342194, -0.15806831419467926, -0.0983254462480545, 0.03375210613012314, 0.1498231291770935, -0.02633300982415676, 0.08807148039340973, -0.041818950325250626, 0.005491909570991993, 0.04332711547613144, -0.09154323488473892, -0.17788946628570557, -0.1511574387550354, 0.03168269246816635, 0.02779703214764595, -0.0036926085595041513, -0.07661224156618118, -0.10547079890966415, -0.02525465004146099, 0.17673082649707794, -0.031458210200071335, -0.0467141754925251, -0.1460992395877838, 0.04803702235221863, 0.14287421107292175, -0.06871066242456436, 0.029714539647102356, 0.001973017817363143, 0.13839562237262726, 0.025180470198392868, -0.1002432331442833, 0.0599750280380249, -0.06703221797943115, -0.1698502004146576, -0.042493224143981934, 0.1219666600227356, 0.05137719213962555, 0.040426451712846756, -0.013905925676226616, 0.019623469561338425, -0.0021664213854819536, -0.0952046811580658, 0.02333015576004982, 0.09779491275548935, 0.06807789951562881, 0.10336580872535706, -0.08477682620286942, 0.01589488610625267, -0.01765877939760685, -0.02125490829348564, 0.14907599985599518, 0.22506071627140045, -0.07158873230218887, 0.09265747666358948, 0.11853967607021332, -0.07960745692253113, -0.17735768854618073, 0.04522618278861046, 0.10486530512571335, 0.016948027536273003, 0.028476761654019356, -0.2281889021396637, 0.10104119032621384, 0.10835017263889313, -0.008374650962650776, 0.0462765134871006, -0.3263278901576996, -0.12231868505477905, 0.07626013457775116, 0.10890765488147736, 0.04243692383170128, -0.13182386755943298, -0.04908346012234688, -0.045424606651067734, -0.1487341821193695, 0.12416041642427444, -0.09715470671653748, 0.11776037514209747, -0.012733484618365765, 0.07528600841760635, 0.026815801858901978, -0.0396973080933094, 0.1431470662355423, 0.00802156887948513, 0.06767663359642029, -0.04821028187870979, 0.02396244741976261, 0.03421349823474884, -0.06809461861848831, 0.052237045019865036, -0.08039446920156479, 0.061707690358161926, -0.13139308989048004, -0.02601500228047371, -0.05496671423316002, 0.07187704741954803, -0.0440305657684803, -0.05018317326903343, -0.054773587733507156, 0.03787750005722046, 0.07670840620994568, -0.021742498502135277, 0.11009840667247772, 0.02569686621427536, 0.08753456175327301, 0.0782877504825592, 0.11002960056066513, -0.013086740858852863, -0.09508194774389267, -0.012708793394267559, -0.03881288319826126, 0.069613978266716, -0.111848384141922, 0.018693363294005394, 0.12236738204956055, 0.032101333141326904, 0.13594259321689606, 0.05302201956510544, -0.06393056362867355, 0.013708047568798065, 0.04927658662199974, -0.1144016832113266, -0.14238381385803223, -0.022200101986527443, -0.03751518204808235, -0.1258607655763626, 0.015106480568647385, 0.11649058759212494, -0.04531902074813843, -0.014201712794601917, -0.013932245783507824, 0.01896754652261734, -0.006588957738131285, 0.15491653978824615, 0.03562546148896217, 0.05100506171584129, -0.08299900591373444, 0.14354893565177917, 0.07514994591474533, -0.10850730538368225, 0.08330834656953812, 0.08537643402814865, -0.0740581750869751, -0.015156284905970097, 0.05831298977136612, 0.16419681906700134, -0.041055869311094284, -0.04780564457178116, -0.09004180878400803, -0.09486015141010284, 0.05823066085577011, 0.08717861771583557, 0.033851467072963715, 0.003838494187220931, -0.021688006818294525, 0.051201026886701584, -0.14195969700813293, 0.07821982353925705, 0.04775024205446243, 0.0690869465470314, -0.13734716176986694, 0.1165732592344284, 0.013676793314516544, 0.03238119184970856, -0.013836224563419819, -0.0026069951709359884, -0.11238475143909454, -0.020509913563728333, -0.14061567187309265, -0.005544288549572229, -0.044500041753053665, 0.012093618512153625, -0.008742650039494038, -0.03176204487681389, -0.04886465519666672, 0.04428255930542946, -0.07262842357158661, -0.06590558588504791, 0.0046875569969415665, 0.06972726434469223, -0.13482928276062012, 0.0016538511263206601, 0.02044961228966713, -0.10762349516153336, 0.0766744539141655, 0.06366874277591705, 0.018507763743400574, 0.036013975739479065, -0.1333671659231186, -0.02531413547694683, 0.026768820360302925, 0.01830645464360714, 0.045573823153972626, -0.08005310595035553, -0.01375824399292469, -0.025205092504620552, 0.047554515302181244, 0.0048302337527275085, 0.04239402711391449, -0.12837590277194977, -0.04437984153628349, -0.07285230606794357, -0.050309330224990845, -0.08137597143650055, 0.05961356684565544, 0.11499307304620743, 0.03600836172699928, 0.16039088368415833, -0.08158952742815018, 0.03707999363541603, -0.1728435903787613, -0.018781086429953575, -0.0033481139689683914, -0.03152909129858017, -0.07127733528614044, -0.029604388400912285, 0.07608474045991898, -0.0632561519742012, 0.13417556881904602, -0.04751462861895561, 0.06116790324449539, 0.033558864146471024, -0.038074713200330734, -0.035374365746974945, 0.01965310424566269, 0.19301870465278625, 0.08251366019248962, -0.018368234857916832, 0.06924780458211899, -0.01443859189748764, 0.06992240995168686, 0.027571506798267365, 0.17022693157196045, 0.1535472869873047, 0.0036635249853134155, 0.08893977850675583, 0.07644268125295639, -0.09237884730100632, -0.13900960981845856, 0.11073105782270432, -0.02048778161406517, 0.11939676851034164, -0.040156226605176926, 0.17610003054141998, 0.08917548507452011, -0.15513946115970612, 0.040491413325071335, -0.03161909431219101, -0.09250882267951965, -0.11444343626499176, -0.04036647453904152, -0.08953369408845901, -0.12744353711605072, 0.025112995877861977, -0.11900734901428223, 0.024249544367194176, 0.04970480501651764, 0.02181127481162548, 0.01911492832005024, 0.11990926414728165, -0.011261544190347195, 0.0010264688171446323, 0.06979845464229584, 0.002975563518702984, -0.006017641164362431, -0.04834635183215141, -0.09718749672174454, 0.05297007039189339, 0.010981619358062744, 0.07960202544927597, -0.031850285828113556, -0.01721300557255745, 0.04890719801187515, -0.018220379948616028, -0.07790879160165787, 0.03383024036884308, 0.015981286764144897, 0.03320213407278061, 0.04315461590886116, 0.04602530971169472, -0.029303384944796562, -0.04764518141746521, 0.2524052858352661, -0.07250767946243286, -0.05361812561750412, -0.12925221025943756, 0.15339291095733643, 0.03348139673471451, -0.003606203244999051, 0.06997084617614746, -0.108570896089077, -0.042133182287216187, 0.15047615766525269, 0.13514412939548492, -0.0467374287545681, -0.018882237374782562, 0.004151931032538414, -0.014514241367578506, -0.0387195385992527, 0.12332215905189514, 0.1002587303519249, 0.053521621972322464, -0.04383575916290283, -0.008984268642961979, -0.006187475752085447, -0.03210844844579697, -0.08601494133472443, 0.05143937095999718, 0.011404324322938919, -0.01049140002578497, -0.03568270802497864, 0.06319660693407059, -0.03740423917770386, -0.15678264200687408, 0.025267615914344788, -0.1448291689157486, -0.16638870537281036, -0.036377694457769394, 0.09066961705684662, -0.007325971499085426, 0.05395810306072235, 0.002596246777102351, -0.023621883243322372, 0.11117178946733475, -0.010956423357129097, -0.06182955577969551, -0.08716527372598648, 0.07150334119796753, -0.014322152361273766, 0.20655560493469238, 0.00827555451542139, 0.07682721316814423, 0.11282742023468018, 0.04172448068857193, -0.12531407177448273, 0.021809274330735207, 0.07268685102462769, -0.07003897428512573, 0.046549178659915924, 0.14998795092105865, -0.05474063381552696, 0.08643312752246857, 0.04187466576695442, -0.10097846388816833, -0.009993058629333973, -0.04678267985582352, -0.02310357615351677, -0.07356339693069458, -0.007160489913076162, -0.06518292427062988, 0.15199466049671173, 0.20473144948482513, -0.033449482172727585, -0.014897161163389683, -0.06280694901943207, 0.03290780261158943, 0.04340201988816261, 0.09386630356311798, -0.030404740944504738, -0.20968040823936462, 0.0074904016219079494, 0.051510997116565704, 0.04660606011748314, -0.24262897670269012, -0.09689858555793762, 0.024124886840581894, -0.03536060079932213, -0.06822656840085983, 0.1085410937666893, 0.07450584322214127, 0.028710786253213882, -0.041410017758607864, -0.11787513643503189, -0.04741530865430832, 0.13942846655845642, -0.15562741458415985, -0.04196322336792946 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # robertagpt2_xsum This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.13.0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "robertagpt2_xsum", "results": []}]}
text2text-generation
Ayham/robertagpt2_xsum
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
# robertagpt2_xsum This model is a fine-tuned version of [](URL on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.13.0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# robertagpt2_xsum\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.13.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n", "# robertagpt2_xsum\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.13.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 53, 31, 6, 12, 8, 3, 118, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n# robertagpt2_xsum\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.13.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.06450962275266647, 0.13180875778198242, -0.0032584774307906628, 0.05736853927373886, 0.1397383064031601, 0.032141927629709244, 0.10620024055242538, 0.12445202469825745, -0.0873434990644455, 0.0898684486746788, 0.08492551743984222, 0.0762886181473732, 0.05597748979926109, 0.11709734797477722, -0.029113251715898514, -0.25106024742126465, 0.0298643559217453, -0.005228882189840078, -0.078975148499012, 0.09030071645975113, 0.10123156011104584, -0.09886722266674042, 0.06883717328310013, 0.007453626953065395, -0.14898230135440826, 0.02297556772828102, -0.05463193356990814, -0.06880732625722885, 0.07899784296751022, 0.022035295143723488, 0.08460967242717743, 0.02059059962630272, 0.08838718384504318, -0.23699086904525757, 0.005850564688444138, 0.06618472933769226, 0.023171167820692062, 0.08279287815093994, 0.07354798167943954, -0.017925245687365532, 0.12110719084739685, -0.14574222266674042, 0.10799827426671982, 0.028016457334160805, -0.06329577416181564, -0.146357923746109, -0.08388671278953552, 0.07640273869037628, 0.07777708768844604, 0.08350301533937454, 0.007210914045572281, 0.1270361840724945, -0.10550515353679657, 0.0778110921382904, 0.1955346018075943, -0.24932940304279327, -0.056512653827667236, 0.01481893751770258, 0.07203095406293869, 0.07190869003534317, -0.09778118878602982, -0.007303302176296711, 0.04006437212228775, 0.036320850253105164, 0.08813733607530594, 0.0014487657463178039, -0.06960494816303253, 0.0036315678153187037, -0.13184861838817596, -0.03525756672024727, 0.14577652513980865, 0.05235500633716583, -0.03893589973449707, -0.125705286860466, -0.03777622804045677, -0.10820334404706955, -0.016717391088604927, -0.03507283329963684, 0.03460036963224411, -0.03769973665475845, -0.09019139409065247, -0.059177741408348083, -0.0635569766163826, -0.05660601705312729, -0.003971700090914965, 0.16485202312469482, 0.05550631880760193, -0.0017879849765449762, -0.03706216812133789, 0.09601981192827225, 0.05340762808918953, -0.11106891185045242, -0.019190745428204536, -0.008870961144566536, -0.07144461572170258, -0.047352246940135956, -0.05915989726781845, -0.06023980677127838, 0.030296290293335915, 0.18333637714385986, -0.05850999057292938, 0.09957373142242432, 0.010393469594419003, 0.007331956643611193, 0.003101900452747941, 0.12198891490697861, -0.036975253373384476, -0.02075781486928463, -0.014650517143309116, 0.09276580810546875, 0.015429477207362652, -0.022165460512042046, -0.08618415147066116, -0.02748715877532959, 0.10093621909618378, 0.06691744178533554, -0.023887312039732933, 0.030098188668489456, -0.05025666952133179, -0.03319039195775986, 0.008724457584321499, -0.12103959172964096, 0.03952464833855629, 0.00511739170178771, -0.07839959859848022, 0.014499910175800323, 0.02695297822356224, 0.00403994508087635, -0.04334385693073273, 0.08536487817764282, -0.07850255072116852, -0.0159909687936306, -0.07289272546768188, -0.06594694405794144, 0.03379519656300545, -0.0699075236916542, -0.0037766864988952875, -0.08371523767709732, -0.15118202567100525, -0.04075537994503975, 0.040650948882102966, -0.050619881600141525, -0.038387905806303024, -0.04967312887310982, -0.043362028896808624, 0.04298538342118263, -0.018473291769623756, 0.13486437499523163, -0.05116066709160805, 0.07080476731061935, -0.013637376949191093, 0.03597098961472511, 0.03472976014018059, 0.045509591698646545, -0.06657436490058899, 0.026233211159706116, -0.11363042891025543, 0.09341259300708771, -0.09822181612253189, 0.016325095668435097, -0.1204552948474884, -0.09131095558404922, -0.012648508884012699, -0.001409029820933938, 0.07523232698440552, 0.12408757209777832, -0.10712955892086029, -0.05099271237850189, 0.1519952118396759, -0.06200286000967026, -0.06938041001558304, 0.09594394266605377, -0.044006768614053726, -0.015552105382084846, 0.05321090668439865, 0.13822096586227417, 0.08943811804056168, -0.10895262658596039, -0.03273468464612961, 0.025881119072437286, 0.054656561464071274, 0.020862936973571777, 0.06989315897226334, -0.018895167857408524, -0.005728050135076046, 0.02213187702000141, -0.025373034179210663, 0.026667410507798195, -0.08633355051279068, -0.0775371789932251, -0.04835537075996399, -0.0756254717707634, 0.05072629451751709, 0.002246761927381158, 0.0411808155477047, -0.07755669206380844, -0.12564323842525482, 0.1108199879527092, 0.13790807127952576, -0.07276078313589096, 0.02318401448428631, -0.071287602186203, 0.020094113424420357, -0.030783023685216904, -0.014580454677343369, -0.20468515157699585, -0.10175057500600815, 0.039045609533786774, -0.09058227390050888, 0.0472310371696949, -0.02063857577741146, 0.06576859205961227, 0.05515943095088005, -0.03151391074061394, -0.04436780884861946, -0.09351322054862976, -0.002434995723888278, -0.09530183672904968, -0.1908087432384491, -0.0506083182990551, -0.03712180256843567, 0.1495523750782013, -0.189472496509552, 0.00937072467058897, 0.0067633637227118015, 0.17014005780220032, 0.029108043760061264, -0.04630865901708603, 0.009877897799015045, 0.05448281392455101, -0.01833798736333847, -0.08036036789417267, 0.0356682650744915, 0.005229397211223841, -0.12982122600078583, -0.03295839577913284, -0.12626972794532776, 0.029701292514801025, 0.09624627977609634, 0.029832616448402405, -0.086390919983387, -0.03733554482460022, -0.06784265488386154, -0.047852255403995514, -0.04929017275571823, -0.006325070746243, 0.18914218246936798, 0.015695571899414062, 0.13055552542209625, -0.056692395359277725, -0.06793022900819778, 0.002643745858222246, 0.007229558192193508, -0.03853912279009819, 0.09774527698755264, 0.07399236410856247, -0.08228262513875961, 0.08558686077594757, 0.09033029526472092, -0.05140385404229164, 0.15323856472969055, -0.04937109351158142, -0.0945889949798584, -0.015191640704870224, 0.03192219138145447, -0.008214701898396015, 0.06115676462650299, -0.11019857227802277, -0.0029652914963662624, 0.030742991715669632, 0.028606586158275604, 0.0568174384534359, -0.1581411212682724, 0.016091711819171906, 0.02395780012011528, -0.033461153507232666, 0.014420620165765285, -0.006838023196905851, 0.02543969266116619, 0.0801130160689354, 0.0485687218606472, -0.00019988560234196484, 0.033674538135528564, -0.007301484700292349, -0.07677152752876282, 0.1854752004146576, -0.13896812498569489, -0.17027391493320465, -0.13482984900474548, 0.03485202416777611, -0.06904032081365585, -0.03251980245113373, 0.029969114810228348, -0.09911543130874634, -0.061586394906044006, -0.0748177170753479, -0.003642365802079439, -0.08534981310367584, 0.009918276220560074, 0.05911427363753319, 0.018550019711256027, 0.058623604476451874, -0.13940121233463287, 0.010055425576865673, -0.013868920505046844, -0.07050018757581711, 0.003617719514295459, 0.05117353796958923, 0.09803489595651627, 0.11502733826637268, -0.000012267176316527184, 0.014063599519431591, -0.03385697677731514, 0.20723828673362732, -0.06799732148647308, -0.032155584543943405, 0.11154734343290329, -0.000450657622423023, 0.06443773210048676, 0.07727853953838348, 0.034726232290267944, -0.08162827044725418, 0.03876209631562233, 0.060257334262132645, -0.006374793592840433, -0.2643311023712158, -0.04351310431957245, -0.037761058658361435, -0.06568510085344315, 0.10303034633398056, 0.06254731118679047, 0.0016801507445052266, 0.05008848384022713, -0.025323444977402687, 0.055742692202329636, -0.02091239206492901, 0.09054747223854065, 0.0995764434337616, 0.042708706110715866, 0.08953706920146942, -0.03381575644016266, -0.03525480628013611, 0.07426388561725616, 0.008714396506547928, 0.24408607184886932, -0.014310845173895359, 0.137417271733284, 0.036665741354227066, 0.11813292652368546, -0.014366279356181622, 0.02621963433921337, 0.02216099016368389, 0.016097402200102806, -0.0024637016467750072, -0.06274297833442688, -0.0394660048186779, 0.0284651480615139, 0.02044367603957653, 0.02779643051326275, -0.11012725532054901, 0.020379047840833664, 0.026383498683571815, 0.2625339925289154, 0.006880008149892092, -0.28376078605651855, -0.07077837735414505, 0.02258957177400589, -0.03919490799307823, -0.06096470355987549, 0.012828348204493523, 0.088597871363163, -0.13607650995254517, 0.06275849044322968, -0.05225657299160957, 0.09273770451545715, -0.06264759600162506, -0.0089786471799016, 0.01867925375699997, 0.15287692844867706, -0.0012540844036266208, 0.08334608376026154, -0.21641774475574493, 0.18204396963119507, 0.016728198155760765, 0.10747934877872467, -0.06565509736537933, 0.0493294782936573, 0.022541845217347145, 0.028714463114738464, 0.08525484800338745, -0.008516172878444195, -0.10011769831180573, -0.1559617519378662, -0.09435713291168213, 0.03331461176276207, 0.14907220005989075, -0.027555091306567192, 0.08845234662294388, -0.04447425529360771, 0.007161865476518869, 0.04238138347864151, -0.08730624616146088, -0.17648151516914368, -0.15147711336612701, 0.03152836859226227, 0.022816171869635582, -0.004420313518494368, -0.07717542350292206, -0.10444773733615875, -0.02789120003581047, 0.17804889380931854, -0.029571183025836945, -0.04936688393354416, -0.14808864891529083, 0.05020968243479729, 0.14049309492111206, -0.06961163133382797, 0.02875140868127346, 0.002925306558609009, 0.13725455105304718, 0.024659860879182816, -0.10225476324558258, 0.06088550388813019, -0.06800241023302078, -0.1679385006427765, -0.04193490743637085, 0.12280833721160889, 0.04969795048236847, 0.03897571191191673, -0.016636407002806664, 0.02024495229125023, -0.003552166046574712, -0.09319819509983063, 0.03062448836863041, 0.10234194248914719, 0.06878684461116791, 0.10441823303699493, -0.08563590794801712, 0.014209490269422531, -0.01746588572859764, -0.023400625213980675, 0.14967086911201477, 0.2257268875837326, -0.07022039592266083, 0.09321589022874832, 0.11312360316514969, -0.08247648924589157, -0.17820842564105988, 0.04568390175700188, 0.10692380368709564, 0.018105633556842804, 0.0359615795314312, -0.2281542271375656, 0.09862247109413147, 0.10862800478935242, -0.008336865343153477, 0.04870260879397392, -0.32786399126052856, -0.12164504826068878, 0.07461889833211899, 0.11172538250684738, 0.04145030677318573, -0.13545078039169312, -0.05006198212504387, -0.04593566060066223, -0.15023060142993927, 0.12064722180366516, -0.09456022083759308, 0.11963178217411041, -0.012837866321206093, 0.07376231253147125, 0.02710275538265705, -0.038851842284202576, 0.1435215175151825, 0.005080349277704954, 0.067705899477005, -0.04767949879169464, 0.024159740656614304, 0.03313519433140755, -0.06748687475919724, 0.05179961398243904, -0.08449335396289825, 0.06000000238418579, -0.13687148690223694, -0.025476185604929924, -0.05455337092280388, 0.07346851378679276, -0.043123215436935425, -0.051171280443668365, -0.05183503031730652, 0.037112027406692505, 0.07638030499219894, -0.02182944491505623, 0.10988148301839828, 0.02485623210668564, 0.088991679251194, 0.08344708383083344, 0.10743916779756546, -0.010754738003015518, -0.10195228457450867, -0.014762814156711102, -0.03803917020559311, 0.06875572353601456, -0.1134675145149231, 0.019954033195972443, 0.12005557864904404, 0.03227812424302101, 0.1387750804424286, 0.05221812427043915, -0.06495367735624313, 0.016006385907530785, 0.04830615967512131, -0.11398319900035858, -0.14593727886676788, -0.02440515160560608, -0.036264561116695404, -0.1258733868598938, 0.015160496346652508, 0.11729112267494202, -0.04235716536641121, -0.01580747775733471, -0.015725666657090187, 0.021977318450808525, -0.006638836581259966, 0.15688762068748474, 0.03348635509610176, 0.05090171471238136, -0.08179699629545212, 0.14575260877609253, 0.07413074374198914, -0.1095995157957077, 0.08576754480600357, 0.08532841503620148, -0.06978598982095718, -0.014728209003806114, 0.0571453720331192, 0.16314876079559326, -0.038987912237644196, -0.048124074935913086, -0.09288981556892395, -0.09541083127260208, 0.058255624026060104, 0.08115967363119125, 0.03347361087799072, 0.0030190562829375267, -0.022307969629764557, 0.05065161734819412, -0.1393667757511139, 0.0809055045247078, 0.04918092489242554, 0.06887268275022507, -0.13892534375190735, 0.11515261232852936, 0.013924900442361832, 0.03241985663771629, -0.013070935383439064, -0.003033619374036789, -0.11194545030593872, -0.021538635715842247, -0.14370237290859222, -0.005958086345344782, -0.04217330738902092, 0.010799036361277103, -0.009150057099759579, -0.03103751502931118, -0.04902389273047447, 0.04634646698832512, -0.07472792267799377, -0.06824740767478943, 0.005254821386188269, 0.07009046524763107, -0.13276325166225433, 0.001138741266913712, 0.021000472828745842, -0.10746531933546066, 0.07642479240894318, 0.06528826057910919, 0.019814733415842056, 0.03430083394050598, -0.1289319545030594, -0.02582377940416336, 0.026705749332904816, 0.017134210094809532, 0.045437537133693695, -0.08303689956665039, -0.013309121131896973, -0.025780027732253075, 0.045552052557468414, 0.004123805556446314, 0.04127190262079239, -0.1313454508781433, -0.041286081075668335, -0.07214563339948654, -0.05018891394138336, -0.08211469650268555, 0.05919935926795006, 0.11399157345294952, 0.03650633245706558, 0.1606101244688034, -0.08180107921361923, 0.03801368921995163, -0.1762295961380005, -0.01809326373040676, -0.0036826010327786207, -0.0336168073117733, -0.07107215374708176, -0.03019687719643116, 0.07485935837030411, -0.06270739436149597, 0.13242238759994507, -0.04078353941440582, 0.061545759439468384, 0.03448612615466118, -0.04150909185409546, -0.030512100085616112, 0.018876081332564354, 0.19442495703697205, 0.07837368547916412, -0.016607586294412613, 0.06686722487211227, -0.014845727942883968, 0.07032562792301178, 0.028115464374423027, 0.17109717428684235, 0.1568002551794052, 0.0032711303792893887, 0.08907210826873779, 0.0754970982670784, -0.09257381409406662, -0.14481255412101746, 0.10902512818574905, -0.016795046627521515, 0.12157227843999863, -0.04208630695939064, 0.17283569276332855, 0.08718014508485794, -0.15606969594955444, 0.04268274083733559, -0.03257035091519356, -0.09055054932832718, -0.11516408622264862, -0.036433763802051544, -0.08939839154481888, -0.1280519664287567, 0.024241510778665543, -0.11879018694162369, 0.026504799723625183, 0.04923262074589729, 0.01923510618507862, 0.01790713332593441, 0.11733186990022659, -0.01445615105330944, -0.0006539305322803557, 0.0689363032579422, 0.0020062753465026617, -0.0070348153822124004, -0.04770832508802414, -0.0960085541009903, 0.054019078612327576, 0.012659790925681591, 0.07925830781459808, -0.031797874718904495, -0.02009398862719536, 0.04783467948436737, -0.019374191761016846, -0.07776486128568649, 0.03250590339303017, 0.016967423260211945, 0.03155431151390076, 0.04479685053229332, 0.04731373116374016, -0.02952471561729908, -0.04807000607252121, 0.2548801898956299, -0.07434536516666412, -0.05063263326883316, -0.12812356650829315, 0.15657593309879303, 0.03382810577750206, -0.002861632965505123, 0.07101249694824219, -0.11103737354278564, -0.04053127020597458, 0.1475420743227005, 0.13512347638607025, -0.04277995973825455, -0.0185712818056345, 0.004135222174227238, -0.01479059923440218, -0.03902248293161392, 0.12364835292100906, 0.09959676116704941, 0.05437140911817551, -0.045499201864004135, -0.0027375291101634502, -0.00531104626134038, -0.032495416700839996, -0.0873844102025032, 0.055690355598926544, 0.011307205073535442, -0.009452964179217815, -0.032475441694259644, 0.062029801309108734, -0.03263412043452263, -0.16021080315113068, 0.024616090580821037, -0.14723750948905945, -0.16798973083496094, -0.0367153100669384, 0.09372914582490921, -0.005445318762212992, 0.055672284215688705, 0.002070098649710417, -0.02167532779276371, 0.11433543264865875, -0.010164332576096058, -0.0621846579015255, -0.08825922757387161, 0.07332020998001099, -0.01940854638814926, 0.2016669064760208, 0.0066641210578382015, 0.07349788397550583, 0.11361224204301834, 0.04005027562379837, -0.12460820376873016, 0.022513387724757195, 0.07208847254514694, -0.0699901357293129, 0.04939776659011841, 0.15247245132923126, -0.05371116101741791, 0.08883775025606155, 0.040487855672836304, -0.09846749156713486, -0.011921233497560024, -0.04775344580411911, -0.02447419799864292, -0.07217281311750412, -0.006507144775241613, -0.06829721480607986, 0.15196146070957184, 0.20643824338912964, -0.03427278995513916, -0.014659715816378593, -0.06434211879968643, 0.03460477292537689, 0.04489650949835777, 0.09372154623270035, -0.03120552934706211, -0.21023744344711304, 0.008089167065918446, 0.05651768669486046, 0.04660504311323166, -0.24673743546009064, -0.09717165678739548, 0.024222254753112793, -0.035435304045677185, -0.06997125595808029, 0.1060972660779953, 0.07689128816127777, 0.028497235849499702, -0.041997477412223816, -0.12148125469684601, -0.04943469911813736, 0.1401224434375763, -0.1520851105451584, -0.03927605226635933 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # robertagpt2_xsum2 This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.13.0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "robertagpt2_xsum2", "results": []}]}
text2text-generation
Ayham/robertagpt2_xsum2
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
# robertagpt2_xsum2 This model is a fine-tuned version of [](URL on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.13.0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# robertagpt2_xsum2\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.13.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n", "# robertagpt2_xsum2\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.13.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 53, 32, 6, 12, 8, 3, 118, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n# robertagpt2_xsum2\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.13.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.06702836602926254, 0.14093177020549774, -0.0031236966606229544, 0.06092473119497299, 0.14234495162963867, 0.03486324101686478, 0.10328124463558197, 0.12492383271455765, -0.09271768480539322, 0.09305500984191895, 0.08534219861030579, 0.07308220863342285, 0.060824763029813766, 0.12273406982421875, -0.027392268180847168, -0.25733286142349243, 0.02973964810371399, -0.008355521596968174, -0.07497164607048035, 0.09097499400377274, 0.09565925598144531, -0.10043682903051376, 0.07355216890573502, 0.010194320231676102, -0.15505248308181763, 0.01934092864394188, -0.05618186295032501, -0.07686375826597214, 0.07515640556812286, 0.014089222066104412, 0.08598432689905167, 0.014016249217092991, 0.09210697561502457, -0.2331327646970749, 0.005091169383376837, 0.06381102651357651, 0.026883626356720924, 0.08751609921455383, 0.08152999728918076, -0.010372334159910679, 0.13073848187923431, -0.14390087127685547, 0.10562647134065628, 0.03019242361187935, -0.06191738694906235, -0.12464167177677155, -0.08177796751260757, 0.08069641888141632, 0.07713643461465836, 0.09089189022779465, 0.009922358207404613, 0.126907616853714, -0.10108944028615952, 0.07556552439928055, 0.19336095452308655, -0.24362298846244812, -0.06048314645886421, 0.012343691661953926, 0.07015404850244522, 0.06583569198846817, -0.10498274117708206, -0.015764744952321053, 0.038328684866428375, 0.037665095180273056, 0.08061155676841736, 0.0018415494123473763, -0.05597086250782013, 0.00037418576539494097, -0.1339775174856186, -0.03373856842517853, 0.1491847187280655, 0.05346749350428581, -0.0366639718413353, -0.12201008200645447, -0.038304612040519714, -0.10770885646343231, -0.019095756113529205, -0.03443059325218201, 0.02820037491619587, -0.037763651460409164, -0.09115836024284363, -0.04899027198553085, -0.06284692883491516, -0.054124534130096436, -0.0009002666338346899, 0.15446656942367554, 0.05264643207192421, 0.0007987734861671925, -0.0415409691631794, 0.09008602052927017, 0.06356452405452728, -0.11652510613203049, -0.014576260931789875, -0.013062785379588604, -0.0659736692905426, -0.050469353795051575, -0.047492582350969315, -0.049190863966941833, 0.033547293394804, 0.1618390679359436, -0.06589066237211227, 0.09795066714286804, 0.013126258738338947, 0.006802585907280445, 0.008137905038893223, 0.12173961848020554, -0.03550036624073982, -0.02963762916624546, -0.013634596951305866, 0.09592438489198685, 0.01409350149333477, -0.019738318398594856, -0.0825921818614006, -0.023817041888833046, 0.10436933487653732, 0.06718435138463974, -0.018213054165244102, 0.023626107722520828, -0.049859073013067245, -0.035525571554899216, 0.009739351458847523, -0.12095600366592407, 0.04277193918824196, 0.0040480270981788635, -0.08026052266359329, 0.016851156949996948, 0.01862739957869053, 0.003243799088522792, -0.04566553980112076, 0.09394354373216629, -0.08400814235210419, -0.01594861038029194, -0.07349270582199097, -0.06190513074398041, 0.03716886788606644, -0.06646651774644852, -0.006644579581916332, -0.07968060672283173, -0.15269798040390015, -0.03979210928082466, 0.04070519655942917, -0.051900751888751984, -0.045337025076150894, -0.05031003803014755, -0.04934397339820862, 0.04163861647248268, -0.013110035099089146, 0.1260765641927719, -0.05111200734972954, 0.07275313884019852, -0.013472634367644787, 0.03532801568508148, 0.03648650273680687, 0.047646716237068176, -0.07209432125091553, 0.02171088382601738, -0.10785001516342163, 0.08805661648511887, -0.09368856996297836, 0.020846087485551834, -0.11902358382940292, -0.09483616054058075, -0.016325846314430237, -0.0075398655608296394, 0.07707744091749191, 0.12760140001773834, -0.1106317862868309, -0.05002358928322792, 0.1444329023361206, -0.0677119791507721, -0.0733882263302803, 0.09049548953771591, -0.040406111627817154, -0.015037901699543, 0.045957013964653015, 0.1380397379398346, 0.08691835403442383, -0.10851934552192688, -0.03601481392979622, 0.018618576228618622, 0.049777526408433914, 0.012955175712704659, 0.07006057351827621, -0.016785364598035812, -0.010383790358901024, 0.022361580282449722, -0.012343032285571098, 0.02720276266336441, -0.08833595365285873, -0.08017885684967041, -0.04849734529852867, -0.07426099479198456, 0.04049728438258171, 0.009112956933677197, 0.03896770253777504, -0.07517127692699432, -0.11690835654735565, 0.10485478490591049, 0.13643500208854675, -0.06457464396953583, 0.018227702006697655, -0.06680461764335632, 0.02471081167459488, -0.0363389290869236, -0.018476948142051697, -0.20211181044578552, -0.10142022371292114, 0.044167641550302505, -0.09310664236545563, 0.044648490846157074, -0.02013051137328148, 0.06285341084003448, 0.06103183701634407, -0.026057207956910133, -0.0476204939186573, -0.0942903459072113, -0.000863994937390089, -0.09176545590162277, -0.19169466197490692, -0.04744723066687584, -0.03556985408067703, 0.1389288306236267, -0.1847820281982422, 0.005131584592163563, -0.00498471362516284, 0.16814492642879486, 0.02640082687139511, -0.04937257990241051, 0.01122477650642395, 0.05703353136777878, -0.017076030373573303, -0.08094935864210129, 0.037289220839738846, 0.009138179942965508, -0.1199166402220726, -0.03970064967870712, -0.12663041055202484, 0.02893294021487236, 0.09574081003665924, 0.03360290825366974, -0.08526548743247986, -0.030775215476751328, -0.07343614101409912, -0.05053175985813141, -0.057401392608881, 0.006273442413657904, 0.18679673969745636, 0.01670841872692108, 0.12640653550624847, -0.05568772926926613, -0.06586578488349915, 0.00951449479907751, 0.010547478683292866, -0.04529450088739395, 0.09419551491737366, 0.08331487327814102, -0.08655180037021637, 0.08820970356464386, 0.09287337213754654, -0.04904988035559654, 0.15643256902694702, -0.04957428202033043, -0.09458786994218826, -0.016521355137228966, 0.02952910214662552, -0.010016846470534801, 0.06656557321548462, -0.10635588318109512, 0.002531864680349827, 0.02787139266729355, 0.028105976060032845, 0.057860735803842545, -0.15569031238555908, 0.013931033201515675, 0.018808918073773384, -0.028228718787431717, 0.010238626971840858, -0.0007535301265306771, 0.026976726949214935, 0.07882364094257355, 0.05183621123433113, -0.000041000570490723476, 0.03195510059595108, -0.0066713253036141396, -0.07738418132066727, 0.18057668209075928, -0.13899299502372742, -0.17745362222194672, -0.13493597507476807, 0.045870617032051086, -0.07223142683506012, -0.034772422164678574, 0.03034306690096855, -0.10062568634748459, -0.061006560921669006, -0.07499666512012482, -0.014543874189257622, -0.09204746037721634, 0.010116968303918839, 0.06487797945737839, 0.021253041923046112, 0.05700824409723282, -0.13456816971302032, 0.00874402653425932, -0.01389564760029316, -0.06728437542915344, 0.00004915011959383264, 0.05442003533244133, 0.09904080629348755, 0.11244324594736099, -0.0030600058380514383, 0.017890043556690216, -0.03240653872489929, 0.21218310296535492, -0.06129572540521622, -0.030111754313111305, 0.11016528308391571, 0.006272153463214636, 0.06909476965665817, 0.08510639518499374, 0.03412371873855591, -0.07602588087320328, 0.039419230073690414, 0.05779176577925682, 0.00009790372860152274, -0.25504109263420105, -0.0480535589158535, -0.03925980627536774, -0.05976688861846924, 0.10619387030601501, 0.06410667300224304, -0.0013478680048137903, 0.05092292279005051, -0.020974013954401016, 0.04998287558555603, -0.029368296265602112, 0.09711845219135284, 0.09218673408031464, 0.04804117977619171, 0.08543392270803452, -0.03119737096130848, -0.03770767152309418, 0.07465554773807526, 0.003350025974214077, 0.2413100153207779, -0.011366444639861584, 0.12159775197505951, 0.031514838337898254, 0.12364485114812851, -0.008284157142043114, 0.03108217753469944, 0.0252403412014246, 0.016619842499494553, 0.0006663223612122238, -0.06400489062070847, -0.04515273496508598, 0.02755013294517994, 0.02163325808942318, 0.03865422308444977, -0.1141057014465332, 0.03434307500720024, 0.02525905705988407, 0.2556529939174652, 0.004764085169881582, -0.28067097067832947, -0.07099565863609314, 0.02068953588604927, -0.040354516357183456, -0.06432852894067764, 0.016302058473229408, 0.08413354307413101, -0.1359425038099289, 0.05801166594028473, -0.055966928601264954, 0.09306657314300537, -0.060769129544496536, -0.009332997724413872, 0.020026762038469315, 0.15507996082305908, -0.0002688353415578604, 0.08392180502414703, -0.20988263189792633, 0.18529954552650452, 0.01554078608751297, 0.1117836982011795, -0.06803550571203232, 0.04918314144015312, 0.022318832576274872, 0.020967120304703712, 0.08716245740652084, -0.007509041111916304, -0.11018853634595871, -0.1555963009595871, -0.09443499892950058, 0.03380861133337021, 0.14548179507255554, -0.020021280273795128, 0.08321266621351242, -0.04629240557551384, 0.006666697561740875, 0.040213312953710556, -0.09612840414047241, -0.18462970852851868, -0.1470668464899063, 0.034815315157175064, 0.025920160114765167, -0.002010605065152049, -0.07671977579593658, -0.10557164996862411, -0.020833374932408333, 0.1779601126909256, -0.027507971972227097, -0.04369166120886803, -0.1512184739112854, 0.052434440702199936, 0.14717857539653778, -0.0694814994931221, 0.02951689250767231, -0.004650110844522715, 0.138345405459404, 0.024588126689195633, -0.10038013756275177, 0.05531321465969086, -0.07174397259950638, -0.16506406664848328, -0.04476136714220047, 0.11901809275150299, 0.04857597127556801, 0.038194429129362106, -0.017566420137882233, 0.023682747036218643, -0.008230938576161861, -0.095147505402565, 0.02709558792412281, 0.10193280875682831, 0.07213184237480164, 0.09839040786027908, -0.08928881585597992, 0.006285983603447676, -0.017071153968572617, -0.02351473644375801, 0.14982382953166962, 0.21781142055988312, -0.07023128122091293, 0.09296484291553497, 0.11335543543100357, -0.08058436959981918, -0.17466644942760468, 0.04350252449512482, 0.10965538024902344, 0.015681101009249687, 0.03107623942196369, -0.22228696942329407, 0.10628698766231537, 0.10564376413822174, -0.009996985085308552, 0.04320888966321945, -0.32275229692459106, -0.11649895459413528, 0.07012023031711578, 0.1147911474108696, 0.03798932209610939, -0.13513153791427612, -0.045385267585515976, -0.042343270033597946, -0.14406076073646545, 0.11934764683246613, -0.07604797184467316, 0.117986761033535, -0.019030194729566574, 0.07921936362981796, 0.02445944771170616, -0.04398864880204201, 0.137278750538826, 0.004411087371408939, 0.062345538288354874, -0.04885115846991539, 0.01428284402936697, 0.03788159415125847, -0.0693332776427269, 0.05375252291560173, -0.0771569311618805, 0.06664617359638214, -0.13830772042274475, -0.0207228846848011, -0.05538881942629814, 0.07728920876979828, -0.0454193651676178, -0.05350765958428383, -0.04566513001918793, 0.04374353587627411, 0.07834461331367493, -0.02140086330473423, 0.10663758218288422, 0.03138110041618347, 0.08967648446559906, 0.07974160462617874, 0.1032315194606781, -0.007410517428070307, -0.0953073501586914, -0.016942866146564484, -0.03759664297103882, 0.06568745523691177, -0.10559755563735962, 0.015570035204291344, 0.12314976006746292, 0.033563774079084396, 0.13422584533691406, 0.054196786135435104, -0.06352900713682175, 0.013657777570188046, 0.05261499807238579, -0.12189167737960815, -0.1357835978269577, -0.02933286316692829, -0.02533324435353279, -0.13253134489059448, 0.012605779804289341, 0.10652105510234833, -0.04480329155921936, -0.015060478821396828, -0.014612135477364063, 0.025293216109275818, -0.006760388612747192, 0.15260182321071625, 0.030494144186377525, 0.051958117634058, -0.07856129854917526, 0.14841598272323608, 0.07353996485471725, -0.11393075436353683, 0.084771528840065, 0.09694252908229828, -0.07307300716638565, -0.013546399772167206, 0.0657312199473381, 0.17042407393455505, -0.04292241483926773, -0.04721521958708763, -0.08803597837686539, -0.09065213799476624, 0.05873969569802284, 0.08470368385314941, 0.03080465830862522, 0.0014218668220564723, -0.027138512581586838, 0.0492561049759388, -0.14723236858844757, 0.08488621562719345, 0.05555827543139458, 0.06784554570913315, -0.14540067315101624, 0.09974104166030884, 0.012556210160255432, 0.03276078775525093, -0.012327875010669231, -0.004314222373068333, -0.11524742841720581, -0.02538478374481201, -0.12902313470840454, -0.006182004231959581, -0.05421173945069313, 0.0094284787774086, -0.00984972808510065, -0.03235968202352524, -0.05074804276227951, 0.045908961445093155, -0.06976903229951859, -0.07123984396457672, 0.0018005876336246729, 0.06999978423118591, -0.13027535378932953, 0.006326854228973389, 0.019778991118073463, -0.10436499118804932, 0.07147316634654999, 0.06737050414085388, 0.02141432650387287, 0.03095399960875511, -0.12374661862850189, -0.024989694356918335, 0.025687813758850098, 0.012754552997648716, 0.05077437311410904, -0.08813250809907913, -0.011027208529412746, -0.02366643212735653, 0.041622988879680634, 0.0021425560116767883, 0.049658775329589844, -0.13428522646427155, -0.042267538607120514, -0.06895472854375839, -0.05672211945056915, -0.0828157514333725, 0.05526341497898102, 0.1250237375497818, 0.025196751579642296, 0.16214749217033386, -0.08106342703104019, 0.044418513774871826, -0.17384345829486847, -0.01721181906759739, -0.004810265265405178, -0.03152947500348091, -0.07234314829111099, -0.028616197407245636, 0.07504447549581528, -0.06152597814798355, 0.12259627878665924, -0.03668905794620514, 0.06419407576322556, 0.03586803749203682, -0.045969605445861816, -0.03532025218009949, 0.01653936319053173, 0.20502567291259766, 0.07978050410747528, -0.019801799207925797, 0.07195930182933807, -0.01448910590261221, 0.06568720936775208, 0.025300338864326477, 0.17310406267642975, 0.15496818721294403, -0.004558053333312273, 0.08705638349056244, 0.07974227517843246, -0.09118515253067017, -0.13784156739711761, 0.11252950876951218, -0.018227161839604378, 0.1206551045179367, -0.041123323142528534, 0.16593804955482483, 0.09016157686710358, -0.15352217853069305, 0.035974640399217606, -0.03366851434111595, -0.090082086622715, -0.1134205013513565, -0.03475925326347351, -0.08853248506784439, -0.12491191178560257, 0.027769463136792183, -0.11917402595281601, 0.028799179941415787, 0.03622962906956673, 0.018673419952392578, 0.012782295234501362, 0.11895556002855301, -0.010293318890035152, 0.003988573793321848, 0.07019398361444473, 0.0070749991573393345, -0.006097902078181505, -0.045649558305740356, -0.09172093868255615, 0.040323782712221146, 0.012381516396999359, 0.07954099774360657, -0.031500838696956635, -0.025667214766144753, 0.05144890770316124, -0.017880922183394432, -0.07782748341560364, 0.03402264043688774, 0.012416752986609936, 0.027774181216955185, 0.04578467458486557, 0.04592837765812874, -0.02757388725876808, -0.04687008261680603, 0.2515445053577423, -0.07723341882228851, -0.054413195699453354, -0.12479692697525024, 0.16677521169185638, 0.03370853513479233, -0.008875677362084389, 0.07039937376976013, -0.11283902823925018, -0.04242721199989319, 0.1494634449481964, 0.137731671333313, -0.038684021681547165, -0.02168148010969162, 0.0073922621086239815, -0.015957538038492203, -0.046163830906152725, 0.1194477528333664, 0.09637027978897095, 0.05775747448205948, -0.046467237174510956, -0.0009565677610225976, -0.014566037803888321, -0.03817960247397423, -0.08846612274646759, 0.051894038915634155, 0.014802161604166031, -0.012937147170305252, -0.032167866826057434, 0.0659746304154396, -0.04225161299109459, -0.14980348944664001, 0.023890404030680656, -0.1452525109052658, -0.1672368049621582, -0.03301350399851799, 0.09956355392932892, -0.00836737547069788, 0.053815700113773346, 0.006071171723306179, -0.02089584805071354, 0.11269020289182663, -0.009143628180027008, -0.06578069180250168, -0.08717164397239685, 0.0645616352558136, -0.023796504363417625, 0.2001848965883255, 0.008100994862616062, 0.06968826055526733, 0.11447084695100784, 0.04626129940152168, -0.13253958523273468, 0.022928040474653244, 0.07445238530635834, -0.0762980654835701, 0.048307448625564575, 0.157495379447937, -0.05103971064090729, 0.08976655453443527, 0.03917911276221275, -0.09849723428487778, -0.016302986070513725, -0.06106242164969444, -0.017031317576766014, -0.07015611231327057, -0.0008588663768023252, -0.06153246760368347, 0.15937212109565735, 0.20251715183258057, -0.034021973609924316, -0.01712922193109989, -0.06564635783433914, 0.03893043473362923, 0.041205987334251404, 0.09403276443481445, -0.030709283426404, -0.2142176479101181, 0.007046678569167852, 0.04693811759352684, 0.04468226432800293, -0.24245291948318481, -0.09521044790744781, 0.0262895580381155, -0.037373628467321396, -0.07422695308923721, 0.10517898947000504, 0.06978258490562439, 0.023662107065320015, -0.03956373408436775, -0.11168704926967621, -0.04882458597421646, 0.14307866990566254, -0.15442317724227905, -0.04094495624303818 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # robertagpt2_xsum4 This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.13.0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "robertagpt2_xsum4", "results": []}]}
text2text-generation
Ayham/robertagpt2_xsum4
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
# robertagpt2_xsum4 This model is a fine-tuned version of [](URL on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.13.0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# robertagpt2_xsum4\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.13.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n", "# robertagpt2_xsum4\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.13.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 53, 32, 6, 12, 8, 3, 118, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n# robertagpt2_xsum4\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.13.0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.06747028231620789, 0.13965904712677002, -0.0030991295352578163, 0.06139880418777466, 0.14278970658779144, 0.03497439622879028, 0.10305009037256241, 0.1249450221657753, -0.0927860364317894, 0.09291483461856842, 0.0856013223528862, 0.07377110421657562, 0.060331106185913086, 0.12272820621728897, -0.027174124494194984, -0.2574314475059509, 0.029746247455477715, -0.00790436565876007, -0.07445001602172852, 0.09119021892547607, 0.09538055956363678, -0.10074859857559204, 0.07404598593711853, 0.010205846279859543, -0.15581102669239044, 0.01935916766524315, -0.05668516457080841, -0.0763002410531044, 0.0753755047917366, 0.013874641619622707, 0.08598800748586655, 0.013780095614492893, 0.09203311800956726, -0.23234698176383972, 0.005281647201627493, 0.0638417899608612, 0.026855560019612312, 0.08742858469486237, 0.08136605471372604, -0.010915490798652172, 0.13095729053020477, -0.1442965418100357, 0.1054706797003746, 0.03037497214972973, -0.061866167932748795, -0.1250259280204773, -0.0812387764453888, 0.08031857013702393, 0.07675599306821823, 0.0910504162311554, 0.009729874320328236, 0.12672553956508636, -0.10123080760240555, 0.07594570517539978, 0.19264763593673706, -0.24402707815170288, -0.0607486218214035, 0.012944363988935947, 0.07034724205732346, 0.06604021787643433, -0.10517717897891998, -0.015253515914082527, 0.03860076144337654, 0.03831079229712486, 0.08082646131515503, 0.0018613779684528708, -0.056104544550180435, 0.00040097240707837045, -0.13403598964214325, -0.032925982028245926, 0.14906032383441925, 0.053589653223752975, -0.036584626883268356, -0.12164146453142166, -0.038238223642110825, -0.1086505725979805, -0.019625714048743248, -0.03404538333415985, 0.0284078661352396, -0.038001470267772675, -0.09138542413711548, -0.04827665537595749, -0.06263232231140137, -0.05387486517429352, -0.0012227853294461966, 0.15483734011650085, 0.0526706799864769, 0.000751500076148659, -0.041640572249889374, 0.08995673060417175, 0.06361159682273865, -0.11643269658088684, -0.014379793778061867, -0.012990759685635567, -0.06571613997220993, -0.050503626465797424, -0.04792723432183266, -0.04842522740364075, 0.03395688906311989, 0.16186296939849854, -0.06623823940753937, 0.09794844686985016, 0.013786804862320423, 0.0065704574808478355, 0.008284473791718483, 0.12150197476148605, -0.036341212689876556, -0.030117696151137352, -0.013531651347875595, 0.09605022519826889, 0.013895546086132526, -0.01969306915998459, -0.08273717015981674, -0.02364170551300049, 0.104224793612957, 0.0670243576169014, -0.018580235540866852, 0.024453286081552505, -0.049651190638542175, -0.035425975918769836, 0.009885245934128761, -0.12069796770811081, 0.042391616851091385, 0.003852279856801033, -0.08028639853000641, 0.017316408455371857, 0.01814151741564274, 0.0028914634604007006, -0.04541075974702835, 0.09386534243822098, -0.08443710952997208, -0.016158374026417732, -0.07370810955762863, -0.06209370866417885, 0.03682992234826088, -0.06663846224546432, -0.006066102534532547, -0.07971075177192688, -0.15283621847629547, -0.040183622390031815, 0.04091053456068039, -0.052145469933748245, -0.04514576494693756, -0.05040648952126503, -0.0498187392950058, 0.04142478480935097, -0.012827936559915543, 0.1269228160381317, -0.051489830017089844, 0.07233260571956635, -0.012977806851267815, 0.035437166690826416, 0.036071643233299255, 0.04767145588994026, -0.07194235175848007, 0.021444840356707573, -0.10830044001340866, 0.08809182792901993, -0.09404683113098145, 0.021662002429366112, -0.11874598264694214, -0.0951291099190712, -0.015980124473571777, -0.007025077007710934, 0.07701849192380905, 0.12755629420280457, -0.11065662652254105, -0.049928151071071625, 0.14427398145198822, -0.06720400601625443, -0.07370835542678833, 0.09065104275941849, -0.040583569556474686, -0.015262349508702755, 0.046237796545028687, 0.13809901475906372, 0.08676091581583023, -0.10825342684984207, -0.036399561911821365, 0.01896560564637184, 0.04968172684311867, 0.012251238338649273, 0.06953907757997513, -0.016296276822686195, -0.010978342033922672, 0.022648679092526436, -0.01167631708085537, 0.027239784598350525, -0.08862825483083725, -0.08003459870815277, -0.04851236566901207, -0.07457060366868973, 0.04001234844326973, 0.009010028094053268, 0.03878228738903999, -0.07570149749517441, -0.11689457297325134, 0.1039792001247406, 0.13663768768310547, -0.0645442008972168, 0.01781575195491314, -0.06686396896839142, 0.025094851851463318, -0.03655417636036873, -0.018326757475733757, -0.2022058516740799, -0.1014728769659996, 0.044072750955820084, -0.0923309326171875, 0.044542402029037476, -0.020211100578308105, 0.06320948898792267, 0.06098104640841484, -0.0261504128575325, -0.04787832126021385, -0.09365439414978027, -0.0008081852574832737, -0.09211105853319168, -0.19231785833835602, -0.047468096017837524, -0.03568809852004051, 0.13847842812538147, -0.18497370183467865, 0.004996886011213064, -0.005259600933641195, 0.16787207126617432, 0.026102444157004356, -0.04895952343940735, 0.011478475295007229, 0.05751516669988632, -0.01681676134467125, -0.0808679461479187, 0.03771491348743439, 0.00891277939081192, -0.11954871565103531, -0.03979578614234924, -0.12616266310214996, 0.02976849488914013, 0.09610812366008759, 0.03275614604353905, -0.08538459986448288, -0.03042353317141533, -0.07332725077867508, -0.050628259778022766, -0.056523486971855164, 0.0065445671789348125, 0.187357097864151, 0.01641758345067501, 0.12612585723400116, -0.055552661418914795, -0.06567084044218063, 0.009614824317395687, 0.01099732331931591, -0.04478508234024048, 0.09457355737686157, 0.08290285617113113, -0.08592614531517029, 0.08827222883701324, 0.09231256693601608, -0.04929669573903084, 0.15646058320999146, -0.04954589903354645, -0.09454742074012756, -0.016256848350167274, 0.028943518176674843, -0.010457204654812813, 0.06690490245819092, -0.10705041885375977, 0.0019919706974178553, 0.027777450159192085, 0.027918130159378052, 0.05826761946082115, -0.15534135699272156, 0.014083617366850376, 0.018369855359196663, -0.02807447500526905, 0.010569820180535316, -0.0003819144330918789, 0.02663036808371544, 0.07894100248813629, 0.05185851827263832, -0.0010605975985527039, 0.03205594792962074, -0.006331913638859987, -0.07675443589687347, 0.18097981810569763, -0.1389351636171341, -0.1766461580991745, -0.134731262922287, 0.04494865611195564, -0.07280848920345306, -0.03499683365225792, 0.030190780758857727, -0.10112380981445312, -0.06101963296532631, -0.07480918616056442, -0.014986006543040276, -0.0922207161784172, 0.010242189280688763, 0.06466744840145111, 0.0208266731351614, 0.05680033192038536, -0.13446705043315887, 0.008743319660425186, -0.014278769493103027, -0.0672391727566719, 0.0002911454066634178, 0.054563216865062714, 0.09935737401247025, 0.1123930811882019, -0.0031811054795980453, 0.018028371036052704, -0.03229368105530739, 0.2128496617078781, -0.06114104017615318, -0.030185183510184288, 0.11016984283924103, 0.006746490485966206, 0.06868883222341537, 0.08524849265813828, 0.03454827517271042, -0.07639247924089432, 0.03924153745174408, 0.05789954215288162, 0.00045698086614720523, -0.25497210025787354, -0.04804225638508797, -0.0391131155192852, -0.06024469435214996, 0.10584484785795212, 0.0641605406999588, -0.001009198371320963, 0.05114138126373291, -0.02043372206389904, 0.050174497067928314, -0.030301762744784355, 0.09679318219423294, 0.09241054207086563, 0.047745320945978165, 0.08585986495018005, -0.031141161918640137, -0.0379687137901783, 0.07438727468252182, 0.0030540397856384516, 0.24166052043437958, -0.011501378379762173, 0.12147223949432373, 0.03180675953626633, 0.12349630147218704, -0.008166392333805561, 0.03117566928267479, 0.02491130493581295, 0.016244428232312202, 0.000636192737147212, -0.06377686560153961, -0.04476907476782799, 0.026720531284809113, 0.02117808535695076, 0.03897365182638168, -0.11408518254756927, 0.03417913615703583, 0.025325486436486244, 0.2557929754257202, 0.004312991630285978, -0.2817441523075104, -0.07112177461385727, 0.019922945648431778, -0.03989473357796669, -0.06458261609077454, 0.016532596200704575, 0.08359414339065552, -0.13555532693862915, 0.05778149887919426, -0.05576556548476219, 0.09313344210386276, -0.0609455443918705, -0.00878345686942339, 0.020754076540470123, 0.1563710719347, -0.0006324063870124519, 0.08396164327859879, -0.21050268411636353, 0.1852341592311859, 0.015670334920287132, 0.11233332753181458, -0.06789238750934601, 0.04874652624130249, 0.022811688482761383, 0.021991027519106865, 0.08702030777931213, -0.007518371567130089, -0.10989957302808762, -0.15562136471271515, -0.09476856142282486, 0.034309547394514084, 0.1453397572040558, -0.01962185464799404, 0.08348935097455978, -0.045843642204999924, 0.0064056990668177605, 0.04007594659924507, -0.09644006937742233, -0.1847556084394455, -0.1467604786157608, 0.03435880318284035, 0.02663755975663662, -0.0018826931482180953, -0.07624422013759613, -0.10581735521554947, -0.021168138831853867, 0.17754726111888885, -0.026373544707894325, -0.04374317079782486, -0.15144774317741394, 0.05175856500864029, 0.14722184836864471, -0.06942896544933319, 0.02933579497039318, -0.004792210645973682, 0.13770413398742676, 0.025045474991202354, -0.1005873903632164, 0.05550701916217804, -0.07213620841503143, -0.16440613567829132, -0.04450131207704544, 0.11856400966644287, 0.048667266964912415, 0.03783915564417839, -0.017757901921868324, 0.023533854633569717, -0.008910370990633965, -0.0953247919678688, 0.02716626040637493, 0.10128166526556015, 0.07224856317043304, 0.09847573190927505, -0.08971485495567322, 0.006134200841188431, -0.0168229341506958, -0.02333333157002926, 0.14924432337284088, 0.21710695326328278, -0.06983757019042969, 0.09191308170557022, 0.11395115405321121, -0.0803695097565651, -0.17463640868663788, 0.04388821870088577, 0.1092694103717804, 0.015651939436793327, 0.030231371521949768, -0.22272349894046783, 0.10702215135097504, 0.1053912341594696, -0.009601758792996407, 0.0443793460726738, -0.3217345178127289, -0.11618103086948395, 0.0706716924905777, 0.11530841886997223, 0.03947899118065834, -0.1351739466190338, -0.045144326984882355, -0.042212530970573425, -0.1436728984117508, 0.11878921091556549, -0.07692466676235199, 0.11832872033119202, -0.01947033405303955, 0.07961321622133255, 0.024198222905397415, -0.043933767825365067, 0.13700608909130096, 0.004862499423325062, 0.06299706548452377, -0.0488106906414032, 0.014724122360348701, 0.03751774877309799, -0.06909611076116562, 0.05367666855454445, -0.07798946648836136, 0.06653150916099548, -0.13830505311489105, -0.0205658208578825, -0.05588890239596367, 0.07744607329368591, -0.04538526386022568, -0.05351756513118744, -0.04529232531785965, 0.043672047555446625, 0.0777127593755722, -0.021377848461270332, 0.1059589833021164, 0.03116947039961815, 0.08994410932064056, 0.07874142378568649, 0.10336405038833618, -0.008446264080703259, -0.0956401601433754, -0.016675671562552452, -0.03788706660270691, 0.06577807664871216, -0.10545707494020462, 0.014788683503866196, 0.12340106070041656, 0.03319569304585457, 0.13384851813316345, 0.05440035089850426, -0.06330201774835587, 0.013137828558683395, 0.05290605500340462, -0.1219303086400032, -0.13519306480884552, -0.029170040041208267, -0.026766888797283173, -0.13243792951107025, 0.013392405584454536, 0.10655925422906876, -0.045022767037153244, -0.015040514059364796, -0.014811594039201736, 0.024958616122603416, -0.006891052704304457, 0.15300321578979492, 0.030816011130809784, 0.051679402589797974, -0.07920073717832565, 0.14825133979320526, 0.0730542242527008, -0.11400429159402847, 0.08473576605319977, 0.09709426760673523, -0.07356049120426178, -0.013875539414584637, 0.06612074375152588, 0.1714184582233429, -0.04274255409836769, -0.04747006669640541, -0.08838531374931335, -0.09069647639989853, 0.05897941812872887, 0.08492869138717651, 0.03115713596343994, 0.0009948043152689934, -0.02718493528664112, 0.049139052629470825, -0.1474473774433136, 0.08476285636425018, 0.05509951338171959, 0.06825221329927444, -0.145464688539505, 0.09934864193201065, 0.012319852598011494, 0.032748423516750336, -0.012441630475223064, -0.00418952526524663, -0.11543778330087662, -0.025488613173365593, -0.12946408987045288, -0.006340851541608572, -0.05407553166151047, 0.009632819332182407, -0.009814285673201084, -0.03208330273628235, -0.05106908082962036, 0.04604968801140785, -0.06995311379432678, -0.07107309997081757, 0.0027513112872838974, 0.07028114050626755, -0.1301315873861313, 0.006486678030341864, 0.019277215003967285, -0.10407787561416626, 0.07152394950389862, 0.0676732286810875, 0.02125006914138794, 0.030916353687644005, -0.12393780052661896, -0.025197917595505714, 0.02556524984538555, 0.01233507227152586, 0.05129961296916008, -0.08763433247804642, -0.010924402624368668, -0.023730864748358727, 0.04201404005289078, 0.0023149761836975813, 0.04953360557556152, -0.13410277664661407, -0.0419599749147892, -0.06879564374685287, -0.05585767328739166, -0.08314444124698639, 0.05548449605703354, 0.12507647275924683, 0.025037642568349838, 0.16225534677505493, -0.0810905247926712, 0.04381818696856499, -0.1739494949579239, -0.01714886911213398, -0.005006530787795782, -0.03224451467394829, -0.07222752273082733, -0.028728101402521133, 0.07530428469181061, -0.06205318495631218, 0.12333417683839798, -0.03631918877363205, 0.06414586305618286, 0.035453081130981445, -0.04572199657559395, -0.035858042538166046, 0.016452914103865623, 0.204870343208313, 0.07956128567457199, -0.019755538552999496, 0.07210087031126022, -0.013715858571231365, 0.06565219163894653, 0.02550867572426796, 0.17269113659858704, 0.1547393798828125, -0.0041130902245640755, 0.0874999463558197, 0.07969232648611069, -0.09123934060335159, -0.13745492696762085, 0.11252357810735703, -0.018223807215690613, 0.12096411734819412, -0.04115771874785423, 0.1656632274389267, 0.08992473781108856, -0.15289296209812164, 0.03616524487733841, -0.03394143655896187, -0.09011818468570709, -0.11300864070653915, -0.0342521108686924, -0.08833073824644089, -0.12525421380996704, 0.027798984199762344, -0.11940042674541473, 0.028293149545788765, 0.03657899424433708, 0.01892605982720852, 0.012487365864217281, 0.11965887993574142, -0.009477685205638409, 0.0045898547396063805, 0.07050057500600815, 0.006727997213602066, -0.006451787892729044, -0.045653074979782104, -0.09176358580589294, 0.04024535045027733, 0.011855113320052624, 0.07927504926919937, -0.0318765826523304, -0.026390669867396355, 0.05183696001768112, -0.017790718004107475, -0.0779707059264183, 0.03409092500805855, 0.012372798286378384, 0.02758246473968029, 0.04607362300157547, 0.04584044963121414, -0.02808007039129734, -0.04675080627202988, 0.2513714134693146, -0.07711691409349442, -0.054058801382780075, -0.12457878142595291, 0.1673308163881302, 0.033479154109954834, -0.009318901225924492, 0.07002995908260345, -0.11221799999475479, -0.042304977774620056, 0.15016737580299377, 0.13849055767059326, -0.039156123995780945, -0.022034551948308945, 0.007343960460275412, -0.016019465401768684, -0.0461873933672905, 0.1197640597820282, 0.09652490168809891, 0.05678577721118927, -0.04660342261195183, -0.0007621780969202518, -0.014566110447049141, -0.03823129087686539, -0.08855697512626648, 0.05118598788976669, 0.015682965517044067, -0.013131949119269848, -0.031891386955976486, 0.06590841710567474, -0.043021172285079956, -0.1495816558599472, 0.02380029670894146, -0.14498336613178253, -0.16732117533683777, -0.033635642379522324, 0.09947224706411362, -0.007637265603989363, 0.054044656455516815, 0.006006499286741018, -0.020711224526166916, 0.11258469521999359, -0.009501648135483265, -0.06554130464792252, -0.0873817428946495, 0.06451265513896942, -0.024080373346805573, 0.19986426830291748, 0.0079104695469141, 0.069683738052845, 0.11480753868818283, 0.04604615271091461, -0.13280123472213745, 0.02304369956254959, 0.07370436191558838, -0.07657544314861298, 0.048569660633802414, 0.15713870525360107, -0.050862520933151245, 0.08877748996019363, 0.03864935040473938, -0.09769468754529953, -0.01617177203297615, -0.061013564467430115, -0.016949370503425598, -0.07019636034965515, -0.0008768807165324688, -0.06133511662483215, 0.15942278504371643, 0.20316660404205322, -0.03393262252211571, -0.017084019258618355, -0.06583375483751297, 0.03887289762496948, 0.04149000719189644, 0.09323594719171524, -0.03140240162611008, -0.2141728401184082, 0.006991807837039232, 0.04784230515360832, 0.044377777725458145, -0.24274303019046783, -0.09468159824609756, 0.026278236880898476, -0.03738171234726906, -0.07430745661258698, 0.10489476472139359, 0.07021401822566986, 0.02434510923922062, -0.03954079747200012, -0.11052960902452469, -0.04892199859023094, 0.14307168126106262, -0.15487822890281677, -0.041345514357089996 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlmroberta_gpt2_summarization_xsum This model is a fine-tuned version of [](https://huggingface.co/) on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["xsum"], "model-index": [{"name": "xlmroberta_gpt2_summarization_xsum", "results": []}]}
text2text-generation
Ayham/xlmroberta_gpt2_summarization_xsum
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:xsum", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us
# xlmroberta_gpt2_summarization_xsum This model is a fine-tuned version of [](URL on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
[ "# xlmroberta_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us \n", "# xlmroberta_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ 59, 37, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us \n# xlmroberta_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ -0.09046109020709991, 0.12679268419742584, -0.002553753089159727, 0.06640023738145828, 0.147850900888443, 0.04552440345287323, 0.10115859657526016, 0.11274819821119308, -0.0816655382514, 0.07917356491088867, 0.07454223930835724, 0.07347462326288223, 0.057134389877319336, 0.1501695215702057, -0.03757326677441597, -0.24152670800685883, 0.01294923946261406, -0.02451852709054947, -0.07805027812719345, 0.10221321880817413, 0.08873415738344193, -0.11606931686401367, 0.07151590287685394, -0.0014260048046708107, -0.17094165086746216, 0.00617084838449955, -0.03487731143832207, -0.045524101704359055, 0.1048104465007782, 0.005478205159306526, 0.09687785059213638, 0.022408900782465935, 0.1362971067428589, -0.2225220501422882, 0.004060813691467047, 0.09098343551158905, 0.03857998922467232, 0.08739086240530014, 0.05496527627110481, -0.005493165459483862, 0.11707214266061783, -0.14887335896492004, 0.10004635155200958, 0.020115282386541367, -0.07611262053251266, -0.11606626957654953, -0.09704416245222092, 0.02728714607656002, 0.07615676522254944, 0.08870646357536316, 0.010194094851613045, 0.1205720603466034, -0.09344019740819931, 0.0753193199634552, 0.18039873242378235, -0.22840912640094757, -0.06219334155321121, 0.033780913800001144, 0.06716760247945786, 0.07256515324115753, -0.10351759195327759, -0.0017215907573699951, 0.023757116869091988, 0.022131850942969322, 0.0967361256480217, -0.00587256345897913, -0.10955806076526642, 0.006070747971534729, -0.12277822941541672, -0.006352447904646397, 0.08624253422021866, 0.02728426083922386, -0.02830403670668602, -0.10445559024810791, -0.059133291244506836, -0.12152823060750961, -0.019571732729673386, -0.025218242779374123, 0.04751773551106453, -0.04500637948513031, -0.07293037325143814, -0.04978872835636139, -0.06139891594648361, -0.060228414833545685, -0.02353181131184101, 0.15552188456058502, 0.024698574095964432, 0.011310996487736702, -0.03459585830569267, 0.10001136362552643, 0.030485212802886963, -0.10958025604486465, 0.0007904921076260507, -0.001844023703597486, -0.1103515625, -0.045052625238895416, -0.05708496272563934, -0.025144020095467567, 0.013059443794190884, 0.14874203503131866, -0.0530344694852829, 0.08820214867591858, 0.026395343244075775, -0.015257401391863823, -0.004506370984017849, 0.15433570742607117, -0.0508778840303421, -0.06403642147779465, -0.01027843076735735, 0.08517669886350632, 0.014394205063581467, -0.02626105211675167, -0.0653182789683342, -0.01040652859956026, 0.06958960741758347, 0.058577656745910645, -0.03916703164577484, 0.02833288162946701, -0.040445148944854736, -0.020246461033821106, -0.004547164775431156, -0.12418386340141296, 0.054893288761377335, -0.002247029449790716, -0.08290854096412659, 0.002787095494568348, 0.006105954758822918, 0.016136491671204567, -0.02448773756623268, 0.12590058147907257, -0.08264809846878052, -0.00035957872751168907, -0.09330013394355774, -0.08348539471626282, 0.018014181405305862, -0.13685239851474762, -0.014558389782905579, -0.03623570129275322, -0.1630396842956543, -0.0513259619474411, 0.06025556102395058, -0.05573897808790207, -0.023446012288331985, -0.04697878658771515, -0.04109587520360947, 0.04031723365187645, -0.011506716720759869, 0.17778299748897552, -0.05726100504398346, 0.07045551389455795, -0.037609390914440155, 0.03828350454568863, -0.0019613848999142647, 0.04618695378303528, -0.06702349334955215, 0.012354828417301178, -0.11160257458686829, 0.08219410479068756, -0.06449349969625473, 0.01335627119988203, -0.11069819331169128, -0.08248528838157654, 0.00011534905934240669, -0.007608113810420036, 0.07786550372838974, 0.11315198987722397, -0.20363274216651917, -0.028991051018238068, 0.10583993792533875, -0.06220567971467972, -0.0553349144756794, 0.05978619307279587, -0.05217964947223663, 0.014969702810049057, 0.05134165287017822, 0.15402992069721222, 0.04672067612409592, -0.12292104214429855, -0.02511427365243435, -0.00013482689973898232, 0.04751252382993698, 0.018365226686000824, 0.039794448763132095, -0.005440698936581612, 0.054548852145671844, 0.008673414587974548, -0.04316546022891998, 0.0029927589930593967, -0.0793294608592987, -0.07490794360637665, -0.038298413157463074, -0.0742935836315155, 0.005476529709994793, 0.016136683523654938, 0.015607146546244621, -0.059144243597984314, -0.10662685334682465, 0.11204293370246887, 0.11815980076789856, -0.07061859965324402, 0.025950849056243896, -0.06503389775753021, -0.014234621077775955, 0.0027866249438375235, -0.025345776230096817, -0.20806986093521118, -0.11607245355844498, 0.024798830971121788, -0.06120111048221588, 0.03325744345784187, -0.026532573625445366, 0.055790975689888, 0.0494413822889328, -0.02855030633509159, -0.002258569933474064, -0.06739315390586853, -0.0018968121148645878, -0.10010789334774017, -0.21817414462566376, -0.0506422333419323, -0.019261090084910393, 0.20230495929718018, -0.19802410900592804, -0.005516922101378441, -0.00895639043301344, 0.1336347460746765, 0.02200692892074585, -0.06558578461408615, -0.006879123859107494, 0.05169478431344032, -0.006178928539156914, -0.09625066816806793, 0.03969037905335426, 0.012641261331737041, -0.09721812605857849, -0.036558352410793304, -0.15615728497505188, -0.01135035790503025, 0.0950518250465393, 0.061201825737953186, -0.08328521996736526, -0.04239926487207413, -0.06322528421878815, -0.03717039152979851, -0.07393828779459, 0.024439172819256783, 0.1804240345954895, 0.023548072203993797, 0.1088489443063736, -0.06219373643398285, -0.07496318221092224, 0.005524991080164909, 0.031195607036352158, 0.00033247482497245073, 0.08420712500810623, 0.13044428825378418, -0.07569620758295059, 0.07642592489719391, 0.0647086352109909, -0.03510301560163498, 0.13437695801258087, -0.029292669147253036, -0.07283302396535873, -0.01195572130382061, -0.015007244423031807, -0.021217217668890953, 0.12127944827079773, -0.08103238046169281, 0.016530251130461693, 0.028482357040047646, 0.028990186750888824, 0.04257126525044441, -0.18499422073364258, 0.009900118224322796, 0.019992433488368988, -0.03601314127445221, -0.05417066812515259, -0.0215191338211298, 0.03548279404640198, 0.08935866504907608, 0.025372959673404694, 0.00985316839069128, 0.010702431201934814, -0.008782315999269485, -0.07861173897981644, 0.18483620882034302, -0.1405821144580841, -0.1536647528409958, -0.08253923058509827, 0.014621704816818237, -0.04481123387813568, -0.04078809544444084, 0.0030406329315155745, -0.09435202926397324, -0.06542748212814331, -0.08742067217826843, -0.02522032894194126, -0.024490438401699066, 0.011569321155548096, 0.04783925414085388, 0.00635519390925765, 0.043467145413160324, -0.13383840024471283, 0.010372769087553024, -0.052560411393642426, -0.06294260919094086, 0.010277914814651012, 0.08428776264190674, 0.09077069163322449, 0.12922528386116028, -0.019936896860599518, 0.019286997616291046, -0.024415554478764534, 0.21584904193878174, -0.08761860430240631, 0.006576946936547756, 0.10840870440006256, -0.009842614643275738, 0.04922151193022728, 0.12103753536939621, 0.037450388073921204, -0.09758073091506958, 0.031188884750008583, 0.07296976447105408, -0.024474380537867546, -0.24253803491592407, -0.051680948585271835, -0.03687982261180878, -0.10338626801967621, 0.09284849464893341, 0.04090723395347595, -0.03911377862095833, 0.04011940211057663, -0.0009166962699964643, 0.02800331450998783, -0.01960277371108532, 0.058400657027959824, 0.08031734824180603, 0.05502066761255264, 0.1030498594045639, -0.019824855029582977, -0.011087431572377682, 0.06970962882041931, -0.0001639725815039128, 0.25585341453552246, -0.034963272511959076, 0.09174378961324692, 0.03283888101577759, 0.10969612747430801, -0.01642717234790325, 0.05977405607700348, 0.011501233093440533, -0.008625291287899017, -0.0032782505732029676, -0.05980943143367767, -0.03263847902417183, 0.017717666923999786, -0.013794421218335629, 0.011391948908567429, -0.08486346155405045, 0.05071724206209183, 0.029522299766540527, 0.245924174785614, 0.014503832906484604, -0.2890230417251587, -0.05073392018675804, -0.005258557852357626, -0.025798626244068146, -0.07108627259731293, -0.012698630802333355, 0.11614315956830978, -0.11790519207715988, 0.07475144416093826, -0.06735406816005707, 0.08608860522508621, -0.038669291883707047, -0.003364883130416274, 0.07366583496332169, 0.1644096076488495, -0.014598309993743896, 0.05342873930931091, -0.2017497718334198, 0.2128055989742279, 0.02288399264216423, 0.11783668398857117, -0.07844293862581253, 0.038460373878479004, 0.014145897701382637, 0.014334429986774921, 0.07257139682769775, 0.0006857763510197401, -0.0915142148733139, -0.14537835121154785, -0.08338024467229843, 0.06004409119486809, 0.15444447100162506, -0.012083979323506355, 0.09349193423986435, -0.04474710673093796, 0.006000916939228773, 0.044427886605262756, -0.08878925442695618, -0.15313299000263214, -0.1484285295009613, 0.027001405134797096, 0.020060745999217033, -0.02380235493183136, -0.05001823976635933, -0.0971015989780426, -0.029705561697483063, 0.15883338451385498, -0.013979309238493443, -0.04921437427401543, -0.15874604880809784, 0.07288810610771179, 0.1672953963279724, -0.05126755312085152, 0.037237949669361115, 0.014908742159605026, 0.12419357895851135, 0.02905172109603882, -0.08504051715135574, 0.05742279067635536, -0.07232038676738739, -0.16329322755336761, -0.05858051776885986, 0.10451579093933105, 0.05928192660212517, 0.042921438813209534, -0.014238870702683926, 0.04323773458600044, -0.010184304788708687, -0.09857341647148132, 0.015610386617481709, 0.09790483117103577, 0.05908431485295296, 0.06094296649098396, -0.08702080696821213, 0.03834407404065132, -0.0149297583848238, -0.025084711611270905, 0.12929505109786987, 0.20036348700523376, -0.07916389405727386, 0.10008272528648376, 0.07329288125038147, -0.08831088989973068, -0.1829063594341278, 0.08650332689285278, 0.11146492511034012, 0.026435611769557, 0.04872208833694458, -0.21607352793216705, 0.1304183453321457, 0.12683993577957153, -0.0037823584862053394, 0.060261502861976624, -0.3036276698112488, -0.13390666246414185, 0.029059389606118202, 0.10492943972349167, 0.015913933515548706, -0.11400808393955231, -0.020237252116203308, -0.044294580817222595, -0.13120509684085846, 0.1498863846063614, -0.0709461197257042, 0.11284080147743225, 0.005454549565911293, 0.08390474319458008, 0.02387801930308342, -0.04258905351161957, 0.12756957113742828, 0.028976116329431534, 0.07023930549621582, -0.03483361005783081, 0.007121111731976271, 0.005122946109622717, -0.061319008469581604, 0.02127571776509285, -0.07440339773893356, 0.05769723653793335, -0.10101485252380371, -0.013444599695503712, -0.06073927879333496, 0.05106315761804581, -0.047002390027046204, -0.05611833557486534, -0.03534883260726929, 0.04727977141737938, 0.07652051001787186, -0.0360993891954422, 0.04952783137559891, 0.0005591799272224307, 0.08354445546865463, 0.04203568398952484, 0.10607428103685379, -0.07560306042432785, -0.05861320719122887, 0.011820893734693527, -0.006150827277451754, 0.044276364147663116, -0.10617567598819733, 0.027270449325442314, 0.1404426395893097, 0.053621575236320496, 0.12259573489427567, 0.05326155945658684, -0.03656803071498871, 0.0022920859046280384, 0.05052144080400467, -0.11328288167715073, -0.11437095701694489, 0.019516680389642715, -0.040352318435907364, -0.12256371974945068, 0.008854616433382034, 0.11074275523424149, -0.027855906635522842, -0.011400079354643822, -0.014442447572946548, 0.031831011176109314, 0.00565282441675663, 0.18657708168029785, 0.013154867105185986, 0.05406339839100838, -0.11082842200994492, 0.12809838354587555, 0.06101992353796959, -0.10523751378059387, 0.054510507732629776, 0.09489315003156662, -0.08784020692110062, 0.004616061691194773, 0.07590118050575256, 0.11759131401777267, -0.04907333850860596, -0.03715651482343674, -0.08873043209314346, -0.10170459002256393, 0.050716742873191833, 0.12768064439296722, 0.027065156027674675, 0.002781826537102461, -0.04272673279047012, 0.031369488686323166, -0.15193819999694824, 0.0634634867310524, 0.052283480763435364, 0.06547524780035019, -0.10628005862236023, 0.12826770544052124, 0.022426972165703773, 0.01426108367741108, -0.013764183968305588, 0.02089589461684227, -0.09324204921722412, -0.0243512112647295, -0.0886990875005722, -0.014919312670826912, -0.03139382600784302, -0.0010635252110660076, -0.008456247858703136, -0.04061164706945419, -0.057776231318712234, 0.03203866630792618, -0.07502010464668274, -0.06332123279571533, -0.006249876692891121, 0.03345026820898056, -0.13576450943946838, 0.006393974646925926, 0.017331894487142563, -0.09762810915708542, 0.07960430532693863, 0.05281612649559975, 0.018594196066260338, 0.035357680171728134, -0.15553158521652222, -0.04445122182369232, 0.023772723972797394, 0.022422563284635544, 0.07712312042713165, -0.09287331998348236, -0.009859943762421608, -0.011504157446324825, 0.05017664283514023, 0.01112283580005169, 0.0559011846780777, -0.11452685296535492, 0.003299858421087265, -0.06966033577919006, -0.04418221488595009, -0.06098533421754837, 0.042845554649829865, 0.10853264480829239, 0.04158712923526764, 0.15868385136127472, -0.06262557208538055, 0.03542226180434227, -0.19021527469158173, -0.0254739448428154, -0.01236820314079523, -0.034067682921886444, -0.08898429572582245, -0.026405707001686096, 0.09485191106796265, -0.04736624285578728, 0.13492655754089355, 0.010782980360090733, 0.07463997602462769, 0.03413889557123184, -0.01901528425514698, -0.05844150483608246, 0.011435621418058872, 0.18503360450267792, 0.07567894458770752, -0.012871677987277508, 0.10203026235103607, 0.006511635147035122, 0.06022363901138306, 0.058252062648534775, 0.24337832629680634, 0.14734035730361938, 0.032498423010110855, 0.07981349527835846, 0.05790771543979645, -0.12590138614177704, -0.12328970432281494, 0.14880086481571198, -0.06600162386894226, 0.11985160410404205, -0.0610194057226181, 0.18907606601715088, 0.05276067182421684, -0.1743660271167755, 0.03756874054670334, -0.038112156093120575, -0.10184143483638763, -0.12978778779506683, 0.0032612518407404423, -0.0839654952287674, -0.11407004296779633, 0.018355242908000946, -0.1285458207130432, 0.06166869401931763, 0.0895581915974617, 0.01573008857667446, 0.03741142526268959, 0.10710125416517258, -0.03592505306005478, 0.0006742937839590013, 0.049315135926008224, 0.037353359162807465, 0.004180136602371931, -0.04709727689623833, -0.08887425065040588, 0.030167171731591225, 0.014731839299201965, 0.0930216908454895, -0.044086337089538574, -0.014308597892522812, 0.04896823689341545, -0.00794924795627594, -0.06587912142276764, 0.021489987149834633, 0.004085715860128403, 0.051106832921504974, 0.03729372099041939, 0.0435347855091095, -0.0005456854705698788, -0.05101270228624344, 0.27189573645591736, -0.06170916557312012, -0.08079281449317932, -0.1289130300283432, 0.19836097955703735, 0.029169265180826187, -0.017604555934667587, 0.06852374970912933, -0.09979556500911713, -0.03963780030608177, 0.16269105672836304, 0.1489870846271515, -0.09145920723676682, -0.026866335421800613, -0.011606531217694283, -0.010273371823132038, -0.0520298071205616, 0.1480594277381897, 0.0976891815662384, 0.06932976841926575, -0.05362936854362488, -0.016328036785125732, -0.023311417549848557, -0.043448127806186676, -0.05586051940917969, 0.05501363053917885, 0.012786089442670345, -0.007770858705043793, -0.027030901983380318, 0.06986244767904282, -0.0076683033257722855, -0.18582582473754883, 0.03782714903354645, -0.182536780834198, -0.1847543716430664, -0.01919093169271946, 0.09361714869737625, -0.026938313618302345, 0.047621771693229675, 0.0006667367415502667, -0.01680532842874527, 0.10982391238212585, -0.012612600810825825, -0.04233746603131294, -0.09886834025382996, 0.08130894601345062, -0.0675545185804367, 0.22331088781356812, -0.01185626070946455, 0.06390877813100815, 0.1009954959154129, 0.0530848391354084, -0.12830694019794464, 0.026230350136756897, 0.06000421568751335, -0.11342986673116684, 0.041065726429224014, 0.1376877725124359, -0.060197848826646805, 0.07072075456380844, 0.046369366347789764, -0.10201004892587662, 0.012269917875528336, -0.06115865707397461, -0.035772863775491714, -0.05723278224468231, -0.024065537378191948, -0.06677799671888351, 0.16773386299610138, 0.21050849556922913, -0.01793290302157402, 0.02101319655776024, -0.07372936606407166, 0.02674667350947857, 0.030094077810645103, 0.12003353238105774, -0.04132509604096413, -0.23657944798469543, 0.02933242917060852, 0.028293244540691376, 0.010552577674388885, -0.18919157981872559, -0.0768105611205101, 0.02361905947327614, -0.05558174103498459, -0.0601327009499073, 0.10522890090942383, 0.05993056297302246, 0.0244605615735054, -0.03633907809853554, -0.11232515424489975, -0.023752616718411446, 0.14821426570415497, -0.153090700507164, -0.039680659770965576 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlmroberta_large_gpt2_summarization_cnndm This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "xlmroberta_large_gpt2_summarization_cnndm", "results": []}]}
text2text-generation
Ayham/xlmroberta_large_gpt2_summarization_cnndm
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# xlmroberta_large_gpt2_summarization_cnndm This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
[ "# xlmroberta_large_gpt2_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# xlmroberta_large_gpt2_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ 62, 44, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# xlmroberta_large_gpt2_summarization_cnndm\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ -0.08907187730073929, 0.15610811114311218, -0.003593545174226165, 0.07492832094430923, 0.13091352581977844, 0.04397495836019516, 0.09891380369663239, 0.1458461582660675, -0.10622286796569824, 0.07420166581869125, 0.08503278344869614, 0.03365070000290871, 0.05701017752289772, 0.1504395455121994, -0.032960645854473114, -0.24292509257793427, 0.009576993994414806, -0.005949670914560556, -0.07077459245920181, 0.10973332822322845, 0.09006579965353012, -0.09242867678403854, 0.07260286808013916, -0.010124683380126953, -0.1296064704656601, 0.01190431322902441, -0.03467988967895508, -0.06104708090424538, 0.08677413314580917, -0.013647018931806087, 0.09111577272415161, 0.02010418474674225, 0.14020027220249176, -0.20520693063735962, -0.0006882242159917951, 0.08241412043571472, 0.038313113152980804, 0.0903041735291481, 0.0584125779569149, -0.010601845569908619, 0.1161714568734169, -0.1800016611814499, 0.0925058051943779, 0.013601712882518768, -0.07731912285089493, -0.0934503898024559, -0.09125185012817383, 0.06677525490522385, 0.09620290249586105, 0.10374395549297333, 0.012805137783288956, 0.10737050324678421, -0.08691222965717316, 0.08682356774806976, 0.16359452903270721, -0.21906401216983795, -0.06325580924749374, 0.08364765346050262, 0.042544055730104446, 0.06946397572755814, -0.09616599977016449, -0.007005459628999233, 0.02521158941090107, 0.025684354826807976, 0.1002899557352066, -0.012882527895271778, -0.08735282719135284, 0.006784406490623951, -0.12321910262107849, -0.04201873764395714, 0.12484482675790787, 0.019313035532832146, -0.013803997077047825, -0.10519739985466003, -0.058966830372810364, -0.14038866758346558, -0.01651511713862419, -0.019611716270446777, 0.03352870047092438, -0.05028500780463219, -0.05504121631383896, -0.06811921298503876, -0.07434218376874924, -0.06545104086399078, -0.007667266298085451, 0.09078256040811539, 0.03310874477028847, 0.015654195100069046, -0.04532032459974289, 0.1275390088558197, 0.04188788682222366, -0.1264541745185852, -0.010636980645358562, 0.007888095453381538, -0.09744014590978622, -0.05500951409339905, -0.024229111149907112, -0.03206982463598251, -0.007802564650774002, 0.1284962296485901, -0.05245000869035721, 0.08345025032758713, 0.014962182380259037, -0.010296131484210491, -0.0248163603246212, 0.14287671446800232, -0.04630138725042343, -0.056690510362386703, -0.0058130607940256596, 0.10709621757268906, 0.004866051487624645, -0.00877817627042532, -0.06814298033714294, -0.02773549221456051, 0.08058984577655792, 0.060645319521427155, -0.03675354644656181, 0.04187421128153801, -0.018807072192430496, -0.02595519833266735, 0.02985534816980362, -0.13568906486034393, 0.045213937759399414, 0.008472553454339504, -0.09271065145730972, -0.0017611472867429256, 0.027728749439120293, -0.011471026577055454, -0.04395917430520058, 0.11752831190824509, -0.07796280831098557, 0.0023172858636826277, -0.07763794809579849, -0.06241592764854431, 0.024278512224555016, -0.09935835748910904, -0.044376224279403687, -0.03318801894783974, -0.22174927592277527, -0.056544117629528046, 0.028784725815057755, -0.054497770965099335, -0.03801959753036499, -0.05979137122631073, -0.07335370779037476, 0.0246892049908638, -0.006913312245160341, 0.15615087747573853, -0.05435677245259285, 0.07412858307361603, -0.021331533789634705, 0.028650177642703056, 0.05490985885262489, 0.03879375755786896, -0.06652893126010895, 0.010084973648190498, -0.12189481407403946, 0.08975308388471603, -0.0767449364066124, 0.006238453555852175, -0.10957753658294678, -0.10194467008113861, -0.0005462619592435658, -0.017085056751966476, 0.04190659895539284, 0.14600230753421783, -0.1920459270477295, -0.024085557088255882, 0.14991381764411926, -0.06439695507287979, -0.012278844602406025, 0.06463389098644257, -0.05550335347652435, 0.005226758308708668, 0.06351979076862335, 0.13436178863048553, 0.09589800983667374, -0.12115675956010818, -0.02126767486333847, -0.0050749327056109905, 0.029520386829972267, 0.008686395362019539, 0.03759892284870148, 0.008586613461375237, 0.05633736774325371, 0.008121277205646038, -0.07307709008455276, -0.008314115926623344, -0.0809769481420517, -0.07381623238325119, -0.051358144730329514, -0.07118294388055801, 0.032684944570064545, 0.020021285861730576, 0.040907278656959534, -0.05490797758102417, -0.10491568595170975, 0.09321941435337067, 0.1348998099565506, -0.06371038407087326, 0.01058538630604744, -0.07198746502399445, -0.025411857292056084, -0.006460480857640505, -0.021847369149327278, -0.19470418989658356, -0.11600986868143082, 0.03317498043179512, -0.05058260262012482, 0.045212168246507645, -0.0008930551703087986, 0.07271705567836761, 0.042071808129549026, -0.044268179684877396, -0.025843126699328423, -0.06246264651417732, -0.009703196585178375, -0.08912446349859238, -0.1772858202457428, -0.0389038510620594, -0.020883161574602127, 0.1655130684375763, -0.23011091351509094, 0.02664380706846714, 0.009762460365891457, 0.15500643849372864, 0.028694242238998413, -0.06216898560523987, 0.028028059750795364, 0.04314090311527252, 0.006553941871970892, -0.09788297116756439, 0.024476736783981323, -0.009486434049904346, -0.07739154249429703, -0.039005063474178314, -0.13089878857135773, -0.0075894794426858425, 0.06532555818557739, 0.0984153300523758, -0.10030412673950195, 0.0004419898323249072, -0.06022411212325096, -0.05651770904660225, -0.09232982993125916, 0.0003316114016342908, 0.1954653263092041, 0.035383667796850204, 0.12631505727767944, -0.051229607313871384, -0.07249689102172852, -0.0019298739498481154, 0.026389027014374733, 0.0008961780113168061, 0.08369304984807968, 0.061303336173295975, -0.06537608802318573, 0.07780784368515015, 0.017259059473872185, -0.056063223630189896, 0.15190064907073975, -0.04708441346883774, -0.08862543106079102, -0.018044447526335716, 0.0005896178190596402, -0.018750349059700966, 0.10369943082332611, -0.04927202686667442, -0.008975528180599213, 0.03071870654821396, 0.02651682123541832, 0.033833060413599014, -0.1717631071805954, 0.001293137320317328, 0.01521823089569807, -0.07150821387767792, -0.011907512322068214, -0.006042834371328354, 0.04670843482017517, 0.09123998880386353, 0.004508219193667173, -0.031428735703229904, 0.018429310992360115, -0.012126817367970943, -0.0791132003068924, 0.1703268438577652, -0.10999634861946106, -0.17748227715492249, -0.12301573157310486, 0.05656988173723221, -0.04217210412025452, -0.04202159866690636, -0.004300187807530165, -0.08312469720840454, -0.05587209761142731, -0.10310006886720657, -0.04187304899096489, -0.01803114265203476, -0.005053970497101545, 0.03589624911546707, 0.012341712601482868, 0.05488614737987518, -0.12945403158664703, 0.012530703097581863, -0.021314723417162895, -0.0804678425192833, 0.014557022601366043, 0.057331591844558716, 0.07027914375066757, 0.13473819196224213, -0.016687126830220222, 0.03185645863413811, -0.025220969691872597, 0.18572911620140076, -0.09028784930706024, 0.02393992990255356, 0.11517997086048126, 0.007228819187730551, 0.052329931408166885, 0.11168108880519867, 0.028099702671170235, -0.06800702959299088, 0.01848040334880352, 0.06845486164093018, -0.019998839125037193, -0.2513117492198944, -0.056008532643318176, -0.025245845317840576, -0.04497754946351051, 0.09817104041576385, 0.06044243276119232, 0.0044728354550898075, 0.02219541370868683, -0.007999828085303307, 0.00811890047043562, -0.010179165750741959, 0.0721919909119606, 0.09277661144733429, 0.046139348298311234, 0.09005444496870041, -0.03041750378906727, -0.026967832818627357, 0.06554892659187317, 0.012417945079505444, 0.23668673634529114, -0.04214121401309967, 0.10123930126428604, 0.0038311202079057693, 0.10314576327800751, -0.01209610141813755, 0.03964134678244591, 0.02314556948840618, 0.005905004218220711, 0.013439686968922615, -0.054205529391765594, -0.017499621957540512, 0.020197810605168343, -0.009001683443784714, 0.006565846037119627, -0.06736980378627777, 0.035050246864557266, 0.016212357208132744, 0.28636470437049866, 0.04125184565782547, -0.2924932837486267, -0.053770698606967926, -0.0012803968274965882, -0.04185329005122185, -0.06165219470858574, -0.008080604486167431, 0.1248665601015091, -0.13814577460289001, 0.06883378326892853, -0.06461194157600403, 0.08351869881153107, -0.08160345256328583, -0.0029447099659591913, 0.05692006275057793, 0.11704578250646591, -0.012419541366398335, 0.06996715813875198, -0.20147480070590973, 0.2103070318698883, 0.018732458353042603, 0.11394242197275162, -0.07082927227020264, 0.02870159037411213, 0.022227386012673378, 0.049903739243745804, 0.09171639382839203, -0.00842549093067646, -0.06139746308326721, -0.13931787014007568, -0.13657666742801666, 0.02617773599922657, 0.11877108365297318, -0.026029568165540695, 0.08430539071559906, -0.023445891216397285, -0.016188204288482666, 0.031033679842948914, -0.14747682213783264, -0.14572037756443024, -0.11055220663547516, 0.04044853895902634, 0.04455830156803131, -0.016494473442435265, -0.060513149946928024, -0.10319872200489044, -0.013662600889801979, 0.17894454300403595, -0.019998114556074142, -0.056522831320762634, -0.14308719336986542, 0.053136587142944336, 0.1534053534269333, -0.048003874719142914, 0.018319997936487198, 0.03565632551908493, 0.12183035910129547, 0.04346443712711334, -0.07746092975139618, 0.024637455120682716, -0.0589129701256752, -0.17771930992603302, -0.055346906185150146, 0.1415204405784607, 0.06374359875917435, 0.047364577651023865, -0.011595265939831734, 0.02964666299521923, 0.006977743003517389, -0.08663865178823471, -0.001598415314219892, 0.09299361705780029, 0.09717653691768646, 0.07200529426336288, -0.07255397737026215, 0.0071243601851165295, -0.036492809653282166, -0.016757367178797722, 0.11758576333522797, 0.16349530220031738, -0.08126260340213776, 0.0955841988325119, 0.04516897350549698, -0.07418400794267654, -0.17038209736347198, 0.05808703601360321, 0.1201845183968544, 0.02847621962428093, 0.040108129382133484, -0.20430873334407806, 0.09768445044755936, 0.12133900076150894, -0.01706216298043728, 0.057656608521938324, -0.34244394302368164, -0.11902846395969391, 0.040648773312568665, 0.07336514443159103, 0.006829339545220137, -0.11604399979114532, -0.038683753460645676, -0.0008966115419752896, -0.12602607905864716, 0.13306866586208344, -0.06626605987548828, 0.11071089655160904, -0.016379961743950844, 0.12321501970291138, 0.02973714843392372, -0.05207176133990288, 0.1272842288017273, 0.06281083822250366, 0.06588684767484665, -0.04157068207859993, 0.005004057195037603, 0.06040751188993454, -0.0819699689745903, 0.0704788863658905, -0.057760827243328094, 0.06797175109386444, -0.1641199290752411, -0.010578108951449394, -0.07790269702672958, 0.04601626098155975, -0.045932475477457047, -0.04134875908493996, -0.0413917638361454, 0.05540333688259125, 0.06260288506746292, -0.03247635439038277, 0.06064427271485329, 0.01037864200770855, 0.06675291061401367, 0.10321897268295288, 0.09830590337514877, -0.004178440198302269, -0.10422481596469879, -0.011844088323414326, -0.008935746736824512, 0.03935131058096886, -0.12597261369228363, 0.020772069692611694, 0.14166906476020813, 0.0457121767103672, 0.14063917100429535, 0.027800418436527252, -0.037673015147447586, -0.01779744401574135, 0.041676707565784454, -0.13219477236270905, -0.09882354736328125, 0.007145837880671024, -0.06515098363161087, -0.1275830715894699, 0.005779091734439135, 0.11488169431686401, -0.0432051345705986, -0.013627858832478523, 0.0035923970863223076, 0.03248405456542969, -0.0011907200096175075, 0.20626530051231384, 0.020398328080773354, 0.06571575999259949, -0.09943528473377228, 0.11346264183521271, 0.07168864458799362, -0.08570358902215958, 0.04492556303739548, 0.11860348284244537, -0.0950642079114914, -0.010457291267812252, 0.06637197732925415, 0.1458338499069214, -0.03397046774625778, -0.033991824835538864, -0.08393916487693787, -0.0832756981253624, 0.063230499625206, 0.11621421575546265, 0.0325980968773365, 0.017068861052393913, -0.05569237098097801, 0.016372481361031532, -0.13883869349956512, 0.07273095101118088, 0.08876276016235352, 0.06311357766389847, -0.11428029090166092, 0.15468265116214752, 0.007148466072976589, 0.03604346886277199, -0.01408798061311245, 0.015198640525341034, -0.06687590479850769, -0.017038779333233833, -0.07762661576271057, 0.0006752104382030666, -0.019814923405647278, -0.009218084625899792, -0.01384732685983181, -0.041068803519010544, -0.04158150404691696, 0.03344058245420456, -0.06764300912618637, -0.05736054852604866, -0.010958509519696236, 0.04791763052344322, -0.12920036911964417, -0.006793550215661526, 0.009230202995240688, -0.09395086020231247, 0.07488956302404404, 0.05664615333080292, 0.006471775472164154, 0.023571599274873734, -0.10008220374584198, -0.014514203183352947, 0.0347016304731369, 0.022410960868000984, 0.05587390810251236, -0.07548335939645767, 0.0026098794769495726, -0.026402894407510757, 0.033130474388599396, 0.025354119017720222, 0.039281152188777924, -0.11210229247808456, 0.009572042152285576, -0.06720240414142609, -0.022872332483530045, -0.0699291005730629, 0.04150447994470596, 0.11810372024774551, 0.03917023167014122, 0.16519100964069366, -0.07669256627559662, 0.044308677315711975, -0.19182778894901276, -0.02113470807671547, -0.00018389523029327393, -0.044391781091690063, -0.05682962387800217, -0.01836714893579483, 0.10260602086782455, -0.052162956446409225, 0.1110350713133812, -0.0005721182096749544, 0.05734716355800629, 0.03483552113175392, -0.03489520400762558, -0.05668225511908531, 0.008204208686947823, 0.1413242369890213, 0.06826165318489075, -0.025225799530744553, 0.11330769956111908, -0.017644818872213364, 0.04707122966647148, 0.04703962802886963, 0.2113797515630722, 0.121553435921669, 0.0014170758659020066, 0.07979631423950195, 0.060042738914489746, -0.11289361119270325, -0.13293851912021637, 0.11280573159456253, -0.06301259249448776, 0.11969946324825287, -0.06017448008060455, 0.1814577430486679, 0.06904232501983643, -0.16691330075263977, 0.048905543982982635, -0.025949224829673767, -0.11281425505876541, -0.10247885435819626, -0.04470295086503029, -0.0800786018371582, -0.09567827731370926, 0.0249676201492548, -0.11928480118513107, 0.06085633859038353, 0.07337448000907898, 0.025919895619153976, 0.012048748321831226, 0.12424338608980179, -0.025701893493533134, 0.0020817567128688097, 0.07462150603532791, 0.025470582768321037, 0.013154219835996628, -0.05173905938863754, -0.06943026185035706, 0.02175159938633442, 0.027317246422171593, 0.09128859639167786, -0.03932599350810051, 0.022328291088342667, 0.03042222373187542, -0.01643829233944416, -0.06681837141513824, 0.022707847878336906, 0.013222502544522285, 0.05326152965426445, 0.06324436515569687, 0.060654085129499435, -0.0048700482584536076, -0.04574070870876312, 0.26763707399368286, -0.07263392955064774, -0.08861927688121796, -0.1333320587873459, 0.18728287518024445, 0.038111522793769836, -0.022382957860827446, 0.06221059337258339, -0.10714289546012878, -0.030641630291938782, 0.15984714031219482, 0.15329872071743011, -0.05491703003644943, -0.018561232835054398, -0.023765450343489647, -0.011095383204519749, -0.015859166160225868, 0.100815050303936, 0.07779332995414734, 0.06542717665433884, -0.047529928386211395, -0.018289312720298767, -0.0025446610525250435, -0.03630966693162918, -0.08367177098989487, 0.05473549664020538, -0.000015735626220703125, 0.0006388062029145658, -0.022777654230594635, 0.07677057385444641, -0.02036226913332939, -0.18344646692276, 0.034237850457429886, -0.17687775194644928, -0.18203842639923096, -0.002284081419929862, 0.07746598869562149, -0.024960409849882126, 0.0418693944811821, 0.0008793625165708363, -0.018544876947999, 0.10800907015800476, -0.029811039566993713, -0.051272016018629074, -0.09728474915027618, 0.06901184469461441, -0.09459299594163895, 0.2167709618806839, 0.001244581420905888, 0.07232961058616638, 0.08793477714061737, 0.014055189676582813, -0.13006888329982758, 0.0394769050180912, 0.05070515349507332, -0.08049240708351135, 0.026569604873657227, 0.17961107194423676, -0.049430981278419495, 0.05853833630681038, 0.030937233939766884, -0.10673884302377701, -0.021021191030740738, -0.031602393835783005, -0.011795728467404842, -0.07884863764047623, -0.02168157882988453, -0.03425245359539986, 0.16074497997760773, 0.2133808732032776, -0.014499352313578129, 0.02313375100493431, -0.06904442608356476, 0.008319602347910404, 0.03314634785056114, 0.08631790429353714, -0.027060968801379204, -0.21253517270088196, 0.017613641917705536, 0.0033656302839517593, 0.03695221245288849, -0.18397319316864014, -0.09077523648738861, 0.019626028835773468, -0.05366811528801918, -0.051555197685956955, 0.11284325271844864, 0.04340289533138275, 0.03307728469371796, -0.013468974269926548, -0.07364674657583237, -0.019152192398905754, 0.14483274519443512, -0.18841670453548431, -0.046823207288980484 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlnet_bert_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "xlnet_bert_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/xlnet_bert_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# xlnet_bert_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
[ "# xlnet_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# xlnet_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ 62, 40, 6, 12, 8, 3, 118, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# xlnet_bert_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ -0.08439991623163223, 0.13756003975868225, -0.003822744358330965, 0.06877533346414566, 0.13706229627132416, 0.03876165300607681, 0.10669495910406113, 0.11600542068481445, -0.1078762486577034, 0.05574687570333481, 0.0719185397028923, 0.06966827809810638, 0.0455457828938961, 0.16003744304180145, -0.029682176187634468, -0.21739239990711212, -0.0006879392894916236, -0.008694575168192387, -0.08204315602779388, 0.11677362024784088, 0.08372819423675537, -0.1025225818157196, 0.057330209761857986, -0.002925894921645522, -0.13319256901741028, 0.022360099479556084, -0.042034488171339035, -0.051198385655879974, 0.09563995152711868, -0.0015311619499698281, 0.09415578842163086, 0.03145190328359604, 0.14558403193950653, -0.20642268657684326, 0.003978490363806486, 0.1017133966088295, 0.03879832103848457, 0.09371630847454071, 0.06007838994264603, -0.015164251439273357, 0.13551796972751617, -0.15393972396850586, 0.10142938792705536, 0.017992762848734856, -0.07612034678459167, -0.12116803228855133, -0.09991853684186935, 0.05682051554322243, 0.08902490139007568, 0.12078730762004852, 0.006682223174721003, 0.1334248185157776, -0.10756011307239532, 0.08218882232904434, 0.177137553691864, -0.20506207644939423, -0.04797593131661415, 0.06172151863574982, 0.02586130052804947, 0.060225699096918106, -0.09403292089700699, -0.01612767204642296, 0.023715417832136154, 0.021063182502985, 0.09922169893980026, 0.003324237186461687, -0.09591171890497208, 0.0002219540037913248, -0.12618987262248993, -0.019282585009932518, 0.07134826481342316, 0.023007523268461227, -0.008147023618221283, -0.09647877514362335, -0.058501631021499634, -0.1444406509399414, -0.023464597761631012, -0.010607549920678139, 0.038360949605703354, -0.05564899742603302, -0.06416018307209015, -0.041681207716464996, -0.06524360179901123, -0.046328354626894, -0.02300175465643406, 0.12500391900539398, 0.04602634534239769, 0.009157566353678703, -0.043927472084760666, 0.11395611613988876, 0.07767876237630844, -0.12161601334810257, 0.012822357006371021, 0.014842632226645947, -0.09674175083637238, -0.049208272248506546, -0.024444594979286194, -0.02049059048295021, -0.008865994401276112, 0.13040868937969208, -0.07498046010732651, 0.0811675488948822, 0.022548649460077286, -0.015112174674868584, -0.017005231231451035, 0.14616933465003967, -0.04613059386610985, -0.04361659288406372, -0.013409540057182312, 0.11072955280542374, 0.006990216206759214, -0.014799949713051319, -0.06806410849094391, 0.010939457453787327, 0.06308674812316895, 0.06540358066558838, -0.03990965336561203, 0.04530388489365578, -0.01919122412800789, -0.017583992332220078, 0.04446640983223915, -0.13802948594093323, 0.044714897871017456, 0.019627943634986877, -0.1038922369480133, -0.0109635591506958, 0.0138162262737751, -0.004990818444639444, -0.050127092748880386, 0.12353028357028961, -0.07622315734624863, 0.009277676232159138, -0.083066426217556, -0.07981564849615097, 0.012794457376003265, -0.09287154674530029, -0.03926530480384827, -0.034623369574546814, -0.17354421317577362, -0.06126809120178223, 0.03083871491253376, -0.05906200408935547, -0.03583678975701332, -0.055912964046001434, -0.07602356374263763, 0.022215334698557854, -0.007946393452584743, 0.18061484396457672, -0.06913577020168304, 0.06506780534982681, -0.02385144680738449, 0.028662484139204025, 0.06800548732280731, 0.04552249610424042, -0.06990069150924683, -0.010581832379102707, -0.0944882184267044, 0.0876837968826294, -0.08299160748720169, 0.005198757164180279, -0.1086961105465889, -0.09312509000301361, 0.027305983006954193, -0.008607432246208191, 0.05421650782227516, 0.1276126503944397, -0.19876863062381744, -0.03604533523321152, 0.112310990691185, -0.05346793681383133, -0.0079732621088624, 0.04976700618863106, -0.05490625277161598, -0.009484664537012577, 0.06251200288534164, 0.11999967694282532, 0.06911466270685196, -0.12484607100486755, -0.006187862250953913, 0.0019448723178356886, 0.034657858312129974, 0.011378062888979912, 0.01511887926608324, 0.007940797135233879, 0.06510118395090103, 0.005467158742249012, -0.03924780339002609, 0.00973390694707632, -0.09532430768013, -0.0682457685470581, -0.033713534474372864, -0.07367049157619476, 0.01708761788904667, 0.0313582606613636, 0.0317513607442379, -0.056407246738672256, -0.11889109760522842, 0.08457308262586594, 0.12642301619052887, -0.05725226551294327, 0.016370821744203568, -0.07644021511077881, -0.03944173827767372, 0.01466435007750988, -0.019018569961190224, -0.19789767265319824, -0.11998051404953003, 0.0203043632209301, -0.06895992904901505, 0.042203716933727264, -0.008983561769127846, 0.06296198070049286, 0.047040048986673355, -0.036310698837041855, -0.0178847536444664, -0.0617828369140625, 0.0023839431814849377, -0.09096255153417587, -0.214678555727005, -0.0344657301902771, -0.01659107767045498, 0.15966935455799103, -0.22380612790584564, 0.01218858826905489, -0.020211581140756607, 0.14470770955085754, 0.024060893803834915, -0.05684486776590347, 0.012893321923911572, 0.05214477702975273, 0.004831715486943722, -0.09477740526199341, 0.03205256909132004, -0.010755632072687149, -0.07749234139919281, -0.015036031603813171, -0.14115989208221436, -0.022937407717108727, 0.08110660314559937, 0.08731198310852051, -0.1020427718758583, 0.022760633379220963, -0.0745810940861702, -0.042460475116968155, -0.08916247636079788, 0.04047457501292229, 0.20772913098335266, 0.035555269569158554, 0.12992805242538452, -0.045981764793395996, -0.08591102808713913, -0.010319259949028492, 0.02292332798242569, 0.011732605285942554, 0.10807662457227707, 0.08167621493339539, -0.04460938647389412, 0.07369378954172134, 0.017556827515363693, -0.041129473596811295, 0.13030804693698883, -0.04426712915301323, -0.08401445299386978, 0.0014501026598736644, -0.02828947827219963, -0.024030614644289017, 0.10024526715278625, -0.03272407874464989, 0.005018276162445545, 0.03101256676018238, 0.02315371297299862, 0.0347837470471859, -0.1867063045501709, 0.0012326734140515327, 0.015333721414208412, -0.05915270745754242, -0.03893366456031799, -0.023093704134225845, 0.05404411628842354, 0.10210999101400375, 0.002182271797209978, -0.03510415926575661, 0.008688711561262608, -0.009006327018141747, -0.07054609805345535, 0.17598596215248108, -0.11579368263483047, -0.17079271376132965, -0.07802461832761765, 0.023759858682751656, -0.03047000989317894, -0.052712660282850266, 0.0007730667712166905, -0.11655708402395248, -0.07318014651536942, -0.1141350194811821, -0.016341688111424446, 0.009922581724822521, -0.002059487160295248, 0.042842477560043335, 0.015428599901497364, 0.05145217105746269, -0.14069320261478424, 0.016182512044906616, -0.053032487630844116, -0.07916940748691559, 0.01679396815598011, 0.08259065449237823, 0.05967662110924721, 0.15351681411266327, -0.015339493751525879, 0.0280049629509449, -0.015772439539432526, 0.19356729090213776, -0.09513691067695618, 0.013467289507389069, 0.09919863939285278, 0.01823655143380165, 0.04305899515748024, 0.12189777195453644, 0.04121058061718941, -0.07653193175792694, 0.027844950556755066, 0.09780647605657578, -0.02197852171957493, -0.2588984966278076, -0.0665772557258606, -0.017334023490548134, -0.08435098081827164, 0.08356811851263046, 0.056824587285518646, -0.009460780769586563, 0.0027056350372731686, -0.01008004043251276, -0.016688073053956032, -0.004118271172046661, 0.04965142905712128, 0.07737123966217041, 0.056620776653289795, 0.09232197701931, -0.032563064247369766, -0.024791117757558823, 0.06293444335460663, -0.003273550420999527, 0.22358781099319458, -0.06509595364332199, 0.07591408491134644, 0.021626228466629982, 0.0906728133559227, -0.013311004266142845, 0.026189981028437614, 0.02369450405240059, -0.014287792146205902, 0.015778018161654472, -0.04815271869301796, 0.0016446862136945128, 0.0014055478386580944, -0.01285536028444767, -0.0031018718145787716, -0.06428773701190948, 0.03740028291940689, 0.015964262187480927, 0.2708616852760315, 0.03559481352567673, -0.2857663333415985, -0.0493142195045948, -0.023348378017544746, -0.03402502089738846, -0.04415789246559143, -0.010588148608803749, 0.09338905662298203, -0.11320807784795761, 0.06551067531108856, -0.04538721591234207, 0.08358901739120483, -0.0754077136516571, -0.002970221685245633, 0.07022926211357117, 0.14651426672935486, -0.01550283469259739, 0.06086865812540054, -0.21200232207775116, 0.2218337506055832, 0.0071553634479641914, 0.12373796850442886, -0.06527459621429443, 0.02137955278158188, 0.01602686010301113, 0.04129594564437866, 0.06097712367773056, -0.005153588484972715, -0.030317990109324455, -0.1566101461648941, -0.11682140827178955, 0.024700891226530075, 0.11401641368865967, 0.0009431575308553874, 0.08326539397239685, -0.02467065304517746, -0.019177183508872986, 0.04112394526600838, -0.15812431275844574, -0.15366724133491516, -0.12621091306209564, 0.03241937980055809, 0.03135211393237114, -0.04297862946987152, -0.05653320252895355, -0.11970193684101105, -0.022664839401841164, 0.1772589087486267, 0.02314941957592964, -0.06138024479150772, -0.1454571634531021, 0.06335586309432983, 0.15080943703651428, -0.043134674429893494, 0.008990980684757233, 0.03630848973989487, 0.1194186881184578, 0.047849297523498535, -0.08706294000148773, 0.038410503417253494, -0.06363960355520248, -0.18691925704479218, -0.056772612035274506, 0.1302841603755951, 0.09158051013946533, 0.049829695373773575, -0.022762442007660866, 0.026915382593870163, 0.001891530817374587, -0.07951032370328903, 0.005784163251519203, 0.08489414304494858, 0.07080216705799103, 0.05920315906405449, -0.06505347043275833, -0.00423869863152504, -0.046100031584501266, -0.021878935396671295, 0.10189201682806015, 0.18754632771015167, -0.07891243696212769, 0.10914573073387146, 0.06561049073934555, -0.0700703114271164, -0.17543338239192963, 0.05727126821875572, 0.12680506706237793, 0.009131605736911297, 0.030789926648139954, -0.20363512635231018, 0.12313258647918701, 0.1173773854970932, -0.010537948459386826, 0.05551096051931381, -0.3556523323059082, -0.12607987225055695, 0.05455269664525986, 0.08338448405265808, 0.0126190809533, -0.10576144605875015, -0.020236026495695114, -0.014705909416079521, -0.12845954298973083, 0.14502887427806854, -0.05860084295272827, 0.10990845412015915, -0.006470322608947754, 0.10267927497625351, 0.023437624797225, -0.04279334098100662, 0.10456269979476929, 0.06211023032665253, 0.07092861831188202, -0.04477464035153389, 0.01908479444682598, 0.036026813089847565, -0.06381380558013916, 0.03573468327522278, -0.0478079654276371, 0.06788234412670135, -0.11709964275360107, -0.013271668925881386, -0.08416878432035446, 0.04521506652235985, -0.0444214753806591, -0.042592473328113556, -0.03997228667140007, 0.0534948855638504, 0.07845401763916016, -0.04604136571288109, 0.06861750781536102, 0.00711180130019784, 0.08586122840642929, 0.06960584968328476, 0.08477172255516052, -0.05356511101126671, -0.09934733808040619, -0.006415478885173798, -0.006482063326984644, 0.033767253160476685, -0.11400053650140762, 0.024955134838819504, 0.13653209805488586, 0.048991501331329346, 0.11978908628225327, 0.03512176126241684, -0.039194557815790176, -0.02172020636498928, 0.03557656705379486, -0.11698087304830551, -0.09105879813432693, 0.03126533702015877, -0.06397838890552521, -0.10354884713888168, 0.014351542107760906, 0.10218588262796402, -0.035822801291942596, -0.018367115408182144, 0.003528810804709792, 0.03191247209906578, 0.014949169009923935, 0.21845993399620056, 0.028139891102910042, 0.0661592110991478, -0.11613281071186066, 0.12655729055404663, 0.06468240171670914, -0.08875774592161179, 0.045512329787015915, 0.133021280169487, -0.09886925667524338, -0.012852138839662075, 0.10359454900026321, 0.13572654128074646, -0.03578856959939003, -0.030785709619522095, -0.09324618428945541, -0.09512593597173691, 0.07871700823307037, 0.13068944215774536, 0.03119000606238842, 0.0020940110553056, -0.05346483364701271, 0.014544796198606491, -0.15040495991706848, 0.0679081603884697, 0.08196306228637695, 0.053497157990932465, -0.08486133068799973, 0.1389232575893402, 0.03018556721508503, 0.03245504945516586, -0.017496975138783455, 0.014001420699059963, -0.0722183957695961, -0.008956614881753922, -0.08571859449148178, -0.012846482917666435, -0.008970885537564754, -0.002487085061147809, -0.021807732060551643, -0.04966174066066742, -0.03998884558677673, 0.04645218700170517, -0.07644140720367432, -0.05997109413146973, -0.0030632512643933296, 0.03841931372880936, -0.13958705961704254, 0.002435486763715744, 0.0021534599363803864, -0.09331252425909042, 0.0680222436785698, 0.0443318672478199, -0.0038094138726592064, 0.03371052071452141, -0.156809002161026, -0.042417336255311966, 0.030696073547005653, 0.018098782747983932, 0.08077108860015869, -0.08009839802980423, -0.001199056743644178, -0.018412325531244278, 0.03907811641693115, 0.012931648641824722, 0.04632226377725601, -0.11373507231473923, 0.005107260774821043, -0.07094312459230423, -0.03258444368839264, -0.06793723255395889, 0.04258469492197037, 0.10688333213329315, 0.04172282665967941, 0.17817431688308716, -0.07975315302610397, 0.03894238546490669, -0.1816628873348236, -0.03182787820696831, -0.003822289640083909, -0.03950243070721626, -0.06582983583211899, -0.01292768307030201, 0.11053968220949173, -0.05359010398387909, 0.12496935576200485, 0.014254624955356121, 0.07255766540765762, 0.041214797645807266, -0.031000765040516853, -0.06443636119365692, 0.015787968412041664, 0.13974204659461975, 0.06263677775859833, -0.01739799976348877, 0.11721600592136383, -0.01281357929110527, 0.0507262721657753, 0.04758094996213913, 0.22862295806407928, 0.127835214138031, 0.03158523514866829, 0.07115469872951508, 0.04858185723423958, -0.14566874504089355, -0.11863180249929428, 0.10679257661104202, -0.09746343642473221, 0.12755714356899261, -0.06721217930316925, 0.20741093158721924, 0.05303938314318657, -0.1441701054573059, 0.055211491882801056, -0.02936212345957756, -0.10340377688407898, -0.10503312200307846, -0.010279136709868908, -0.07648930698633194, -0.11092041432857513, 0.03462722525000572, -0.11014685779809952, 0.06649668514728546, 0.0949074774980545, 0.030536968261003494, 0.02746526524424553, 0.10953677445650101, -0.0015785668510943651, -0.002105593215674162, 0.0560697466135025, 0.02614385448396206, -0.004255656618624926, -0.06553210318088531, -0.0615391731262207, 0.023531120270490646, 0.02372319996356964, 0.08303875476121902, -0.034697629511356354, -0.00787281058728695, 0.049592092633247375, -0.0183393694460392, -0.07462149858474731, 0.03921779617667198, 0.00006163475336506963, 0.06765517592430115, 0.05751412734389305, 0.05671480670571327, 0.006803634110838175, -0.04874961078166962, 0.28759968280792236, -0.06294582039117813, -0.11238265782594681, -0.12273989617824554, 0.2370980679988861, 0.056297630071640015, -0.022878773510456085, 0.061581823974847794, -0.10368084162473679, -0.04950064420700073, 0.15730294585227966, 0.17242412269115448, -0.058964040130376816, -0.027121353894472122, -0.03243726119399071, -0.013212978839874268, -0.030946969985961914, 0.1280723214149475, 0.11382246017456055, 0.06068207696080208, -0.04817952588200569, -0.023585550487041473, -0.019164521247148514, -0.04057344049215317, -0.07270842790603638, 0.07355180382728577, 0.020998746156692505, -0.002751597436144948, -0.0359501913189888, 0.06938734650611877, -0.013004681095480919, -0.21190813183784485, 0.0386723130941391, -0.1687794029712677, -0.18593141436576843, -0.00852117594331503, 0.10152824968099594, -0.023137981072068214, 0.04834151268005371, 0.018530916422605515, -0.015373671427369118, 0.10281042754650116, -0.03001723811030388, -0.03548314422369003, -0.10730046033859253, 0.06429620832204819, -0.13351041078567505, 0.23147282004356384, -0.008416783064603806, 0.06801354140043259, 0.09308183193206787, 0.021483739838004112, -0.126320481300354, 0.0640304908156395, 0.047069743275642395, -0.09982531517744064, 0.03395232930779457, 0.16118015348911285, -0.04167235642671585, 0.04710163548588753, 0.03900695964694023, -0.11526238173246384, -0.0018742805114015937, -0.09286903589963913, -0.030987456440925598, -0.06000611558556557, -0.01930495910346508, -0.02810920961201191, 0.1481609344482422, 0.2165602147579193, -0.009364882484078407, 0.029988881200551987, -0.08095812797546387, -0.0070569273084402084, 0.02114017680287361, 0.08625271916389465, -0.02706008031964302, -0.22401250898838043, 0.03540153056383133, 0.008621535263955593, 0.03932260349392891, -0.1754380464553833, -0.07138241082429886, 0.006853714119642973, -0.07275998592376709, -0.051220089197158813, 0.109926238656044, 0.04716916009783745, 0.04094773530960083, -0.025169147178530693, -0.07999075949192047, -0.009645414538681507, 0.14695115387439728, -0.18950515985488892, -0.03654700517654419 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlnet_distilgpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "xlnet_distilgpt2_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/xlnet_distilgpt2_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# xlnet_distilgpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
[ "# xlnet_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# xlnet_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ 62, 44, 6, 12, 8, 3, 118, 4, 35 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# xlnet_distilgpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.16.2\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.11.0" ]
[ -0.08502685278654099, 0.15221475064754486, -0.0036149013321846724, 0.07184699177742004, 0.12786412239074707, 0.04105905070900917, 0.11117979884147644, 0.13460198044776917, -0.10527464747428894, 0.06351681798696518, 0.08250189572572708, 0.04398773983120918, 0.054805293679237366, 0.16219216585159302, -0.029705902561545372, -0.22787539660930634, 0.007759073283523321, -0.010261399671435356, -0.06414495408535004, 0.11887217313051224, 0.0877949595451355, -0.09133048355579376, 0.06130057945847511, -0.00978323258459568, -0.1287870854139328, 0.020120253786444664, -0.041696976870298386, -0.05937733128666878, 0.08983036130666733, -0.015313222073018551, 0.09095598012208939, 0.030247807502746582, 0.1414242535829544, -0.20931769907474518, 0.00010245395969832316, 0.08522172272205353, 0.03594173863530159, 0.08801662921905518, 0.063563272356987, -0.018881268799304962, 0.11509397625923157, -0.17695926129817963, 0.09691967815160751, 0.013339831493794918, -0.0785403922200203, -0.09351775795221329, -0.09996798634529114, 0.07738246768712997, 0.09653998911380768, 0.11339180916547775, 0.005946803838014603, 0.1258586347103119, -0.10206390917301178, 0.08516886830329895, 0.17355088889598846, -0.2138891965150833, -0.05715882033109665, 0.06390397250652313, 0.02287445217370987, 0.06481114029884338, -0.09005499631166458, -0.014813827350735664, 0.02970052696764469, 0.022842947393655777, 0.09705185145139694, 0.0015595410950481892, -0.09677755832672119, 0.0037081334739923477, -0.12656454741954803, -0.04270966351032257, 0.11870484799146652, 0.021779974922537804, -0.01530565693974495, -0.10782182961702347, -0.04878978058695793, -0.14421382546424866, -0.012467584572732449, -0.012750807218253613, 0.03577464073896408, -0.06101122125983238, -0.05669613555073738, -0.055660247802734375, -0.07768335938453674, -0.04696843400597572, -0.009217008017003536, 0.08822470158338547, 0.04205995053052902, 0.014758163131773472, -0.04444931447505951, 0.123897485435009, 0.06665720790624619, -0.12529972195625305, -0.0025458901654928923, 0.012283879332244396, -0.08955180644989014, -0.05417490005493164, -0.017951687797904015, -0.021090175956487656, -0.0090861851349473, 0.13227714598178864, -0.05812281370162964, 0.07539494335651398, 0.019658485427498817, -0.010480906814336777, -0.01868085376918316, 0.13685715198516846, -0.03398199751973152, -0.049143485724925995, -0.015148652717471123, 0.11562009900808334, 0.004069550894200802, -0.009268154390156269, -0.06736312061548233, -0.0036889042239636183, 0.07835445553064346, 0.06800388544797897, -0.030973847955465317, 0.04242773354053497, -0.01741817779839039, -0.022769784554839134, 0.03517591580748558, -0.14672070741653442, 0.03880387544631958, 0.01120529230684042, -0.10120593756437302, -0.0037790730129927397, 0.02398575283586979, -0.011415978893637657, -0.0568387433886528, 0.10904590785503387, -0.0732477456331253, 0.006967304740101099, -0.07676441222429276, -0.06688669323921204, 0.023793259635567665, -0.10497332364320755, -0.047006864100694656, -0.03413939103484154, -0.20609503984451294, -0.06485055387020111, 0.021527066826820374, -0.052689701318740845, -0.049760960042476654, -0.06365396082401276, -0.08138380199670792, 0.017858123406767845, -0.008178072050213814, 0.1640768051147461, -0.05932379513978958, 0.068272165954113, -0.022735781967639923, 0.024909408763051033, 0.07622469961643219, 0.04157768189907074, -0.06646054983139038, 0.0029549915343523026, -0.0956607386469841, 0.09140218794345856, -0.08536244928836823, 0.006046765949577093, -0.10909666866064072, -0.1003984585404396, 0.019437626004219055, -0.016946356743574142, 0.04026370495557785, 0.14753593504428864, -0.1907208412885666, -0.03001043200492859, 0.13774840533733368, -0.06440536677837372, -0.008140013553202152, 0.06189536303281784, -0.05330081656575203, -0.0010335389524698257, 0.06391679495573044, 0.13243341445922852, 0.08049540966749191, -0.12418682873249054, -0.018242910504341125, -0.0024480249267071486, 0.03691982850432396, 0.009719722904264927, 0.02955687791109085, 0.010029470548033714, 0.06657852232456207, 0.009864733554422855, -0.0705624520778656, -0.003259665332734585, -0.09144172817468643, -0.07459302246570587, -0.04262647405266762, -0.07562936097383499, 0.03578956425189972, 0.02320333756506443, 0.04071814566850662, -0.05945786088705063, -0.10849280655384064, 0.09035126119852066, 0.12874633073806763, -0.05608733370900154, 0.007379766087979078, -0.0775299221277237, -0.03366899490356445, 0.01122364867478609, -0.022792182862758636, -0.19297900795936584, -0.12732252478599548, 0.02838660031557083, -0.05428092181682587, 0.045188434422016144, -0.007077973335981369, 0.06465431302785873, 0.037699341773986816, -0.03826256841421127, -0.02665681205689907, -0.0682259202003479, -0.0053888652473688126, -0.08719377219676971, -0.1884080320596695, -0.048319388180971146, -0.017758361995220184, 0.17189304530620575, -0.22901973128318787, 0.02933388203382492, -0.0024125222116708755, 0.1481030434370041, 0.023186564445495605, -0.06234603747725487, 0.025163860991597176, 0.04993196576833725, 0.007552398834377527, -0.09488997608423233, 0.02536868117749691, -0.010700306855142117, -0.07176204025745392, -0.02373540960252285, -0.13246367871761322, -0.018384169787168503, 0.06899967789649963, 0.10447181016206741, -0.11017842590808868, 0.016384411603212357, -0.07185418903827667, -0.05237008258700371, -0.10138633102178574, 0.016401933506131172, 0.19927076995372772, 0.03585498407483101, 0.12704730033874512, -0.0473211295902729, -0.08699985593557358, -0.006683154497295618, 0.01693984866142273, 0.0015581791521981359, 0.10059439390897751, 0.05783821642398834, -0.05780825763940811, 0.08301516622304916, 0.010124366730451584, -0.04958043619990349, 0.14285194873809814, -0.054053355008363724, -0.08900872617959976, -0.009274494834244251, -0.005842527374625206, -0.020708920434117317, 0.10056126117706299, -0.04099211469292641, -0.015320567414164543, 0.03270285949110985, 0.019892776384949684, 0.03478632867336273, -0.17873340845108032, -0.005703526549041271, 0.015321984887123108, -0.06786885857582092, -0.023242484778165817, -0.023443901911377907, 0.05397465080022812, 0.09806343913078308, -0.0010023885406553745, -0.028908392414450645, 0.018173055723309517, -0.013026612810790539, -0.08367366343736649, 0.1716330498456955, -0.11023394763469696, -0.17718826234340668, -0.10709811747074127, 0.0549260675907135, -0.03630458563566208, -0.053723033517599106, -0.000004516055469139246, -0.1002437099814415, -0.06056733429431915, -0.11458266526460648, -0.03439076617360115, -0.005817305762320757, -0.011060647666454315, 0.04091925173997879, 0.019000932574272156, 0.05113453418016434, -0.13190148770809174, 0.014160952530801296, -0.031469039618968964, -0.08231481164693832, 0.02281995862722397, 0.05916736274957657, 0.06342997401952744, 0.1404442936182022, -0.01570623554289341, 0.03244005888700485, -0.02199958637356758, 0.18993127346038818, -0.0939146876335144, 0.0168188214302063, 0.10859978199005127, 0.01455714926123619, 0.04998750239610672, 0.10618802905082703, 0.03066202998161316, -0.06371970474720001, 0.02053438127040863, 0.08177470415830612, -0.02344430610537529, -0.2670363485813141, -0.06038869917392731, -0.015728725120425224, -0.05527869611978531, 0.09480773657560349, 0.06273268908262253, 0.008341341279447079, 0.012968524359166622, -0.015268170274794102, -0.014265375211834908, -0.004766436293721199, 0.06170206144452095, 0.08285446465015411, 0.04512806609272957, 0.08728466928005219, -0.03418324515223503, -0.026977870613336563, 0.061888981610536575, 0.009573480114340782, 0.23522774875164032, -0.050614990293979645, 0.09258820861577988, 0.012038817629218102, 0.10939986258745193, -0.01338711567223072, 0.03434000909328461, 0.03329981118440628, 0.0006294495542533696, 0.021367821842432022, -0.04946349933743477, -0.004884410183876753, 0.012628882192075253, -0.008267100900411606, 0.001733349054120481, -0.06381331384181976, 0.034592241048812866, 0.014695650897920132, 0.3000965118408203, 0.04017365723848343, -0.28889915347099304, -0.05316714569926262, -0.012677368707954884, -0.04594910144805908, -0.04724933207035065, -0.013408337719738483, 0.10988400876522064, -0.13894535601139069, 0.06919653713703156, -0.04720897227525711, 0.08465225249528885, -0.0778692364692688, -0.004998510703444481, 0.05503404513001442, 0.10840626060962677, -0.01147291250526905, 0.06914553791284561, -0.20151059329509735, 0.22457782924175262, 0.00944904237985611, 0.1162291020154953, -0.07218977063894272, 0.025679849088191986, 0.017883753404021263, 0.04212173819541931, 0.08883678913116455, -0.004632971715182066, -0.04043745994567871, -0.14799971878528595, -0.13623179495334625, 0.01804487593472004, 0.11288005113601685, -0.0243974719196558, 0.08643101900815964, -0.016001148149371147, -0.0171029195189476, 0.026384370401501656, -0.14963461458683014, -0.1450624018907547, -0.1206185445189476, 0.03857434168457985, 0.03448439761996269, -0.02898746356368065, -0.05940726399421692, -0.11706683039665222, -0.020732933655381203, 0.1886610984802246, -0.011072169989347458, -0.06646566092967987, -0.14471934735774994, 0.056561946868896484, 0.1473860740661621, -0.04279955103993416, 0.00803242065012455, 0.04242095723748207, 0.12312673777341843, 0.038142092525959015, -0.0818977952003479, 0.017063111066818237, -0.060204681009054184, -0.18659017980098724, -0.04916601628065109, 0.1422029286623001, 0.0812302678823471, 0.051796332001686096, -0.015347352251410484, 0.02617838978767395, 0.00841572880744934, -0.07784777134656906, 0.0050264280289411545, 0.08829335123300552, 0.08641212433576584, 0.06760640442371368, -0.0677885189652443, -0.007603704929351807, -0.05521763861179352, -0.02253442257642746, 0.11092876642942429, 0.16831833124160767, -0.08254692703485489, 0.10427851974964142, 0.04509342461824417, -0.0750012919306755, -0.167192280292511, 0.04919708892703056, 0.1336037963628769, 0.019811712205410004, 0.038711581379175186, -0.2025069147348404, 0.10819301009178162, 0.1266314834356308, -0.014488968066871166, 0.04795347526669502, -0.35125669836997986, -0.1201833114027977, 0.04903758689761162, 0.06857945770025253, 0.004380464553833008, -0.11507424712181091, -0.03236396238207817, 0.006190500222146511, -0.12409911304712296, 0.13657736778259277, -0.05511550232768059, 0.10374202579259872, -0.009278015233576298, 0.12236732244491577, 0.027305390685796738, -0.04753489792346954, 0.12252172082662582, 0.07994547486305237, 0.06342501193284988, -0.04554011672735214, 0.004704378079622984, 0.05719158425927162, -0.07832116633653641, 0.0705963522195816, -0.0459621287882328, 0.06507615745067596, -0.15663078427314758, -0.011614187620580196, -0.08425629884004593, 0.047051090747117996, -0.04742109775543213, -0.03509264439344406, -0.031895462423563004, 0.055528074502944946, 0.06881241500377655, -0.03811810910701752, 0.0646597221493721, 0.009645801968872547, 0.0879916325211525, 0.09724385291337967, 0.09265730530023575, -0.0158938430249691, -0.116217702627182, -0.018346836790442467, -0.01086526270955801, 0.03216332942247391, -0.11246076971292496, 0.012747562490403652, 0.13856084644794464, 0.05624141916632652, 0.1339241862297058, 0.025895576924085617, -0.044542986899614334, -0.019123602658510208, 0.03448231518268585, -0.12802371382713318, -0.0972789004445076, 0.012310286052525043, -0.05766037106513977, -0.12293965369462967, 0.006548028904944658, 0.1057700514793396, -0.042228929698467255, -0.022288424894213676, -0.00005636271453113295, 0.034362152218818665, 0.009830480441451073, 0.21183234453201294, 0.026552315801382065, 0.0745268240571022, -0.1062701940536499, 0.11476965993642807, 0.07192527502775192, -0.09467103332281113, 0.048693399876356125, 0.1334279626607895, -0.09448616951704025, -0.01574842259287834, 0.09090852737426758, 0.13325318694114685, -0.03540193289518356, -0.03413671255111694, -0.08826032280921936, -0.09351031482219696, 0.07413173466920853, 0.1273440569639206, 0.03653029724955559, 0.015205102041363716, -0.045660994946956635, 0.008052035234868526, -0.137626513838768, 0.0706172063946724, 0.09526905417442322, 0.05607021227478981, -0.09711863845586777, 0.15171818435192108, 0.02136675827205181, 0.04482363536953926, -0.01263174507766962, 0.01012413576245308, -0.0646054819226265, -0.008893813006579876, -0.06808021664619446, 0.0009301693062298, -0.016887599602341652, -0.006639151368290186, -0.015180128626525402, -0.044714365154504776, -0.03462238237261772, 0.041817907243967056, -0.07358632236719131, -0.05971744284033775, -0.007920151576399803, 0.048688411712646484, -0.13258785009384155, -0.0011952577624469995, 0.006150851957499981, -0.10005879402160645, 0.06915630400180817, 0.05534705892205238, -0.002676230389624834, 0.024926062673330307, -0.11056365817785263, -0.02365119941532612, 0.02904190681874752, 0.021464046090841293, 0.06513401865959167, -0.07439400255680084, -0.0007026487728580832, -0.02215626835823059, 0.029325466603040695, 0.027065135538578033, 0.03882871940732002, -0.11388936638832092, 0.007034349720925093, -0.0610857754945755, -0.029289672151207924, -0.07480189204216003, 0.043761104345321655, 0.1162467822432518, 0.03395742177963257, 0.1674564629793167, -0.08070171624422073, 0.04787469655275345, -0.18818189203739166, -0.028312677517533302, -0.002190512837842107, -0.03883690759539604, -0.053640078753232956, -0.012113528326153755, 0.10770726948976517, -0.05087381601333618, 0.11191705614328384, 0.006266195327043533, 0.06303688138723373, 0.038621459156274796, -0.03419985622167587, -0.062161512672901154, 0.014904734678566456, 0.1305404007434845, 0.05969024449586868, -0.021725857630372047, 0.11179914325475693, -0.02794523350894451, 0.03396237641572952, 0.03433287888765335, 0.2188050001859665, 0.12412916868925095, 0.009938668459653854, 0.07523248344659805, 0.0605425126850605, -0.12774042785167694, -0.12475163489580154, 0.09609194099903107, -0.077449269592762, 0.12477254122495651, -0.06477648764848709, 0.18967470526695251, 0.06587062031030655, -0.1583324819803238, 0.05712858587503433, -0.02475224994122982, -0.10751628130674362, -0.10051897913217545, -0.041722409427165985, -0.07475171238183975, -0.09374631196260452, 0.031826891005039215, -0.11740680038928986, 0.06218937784433365, 0.09153898805379868, 0.030238449573516846, 0.019534720107913017, 0.11177211999893188, -0.0106222378090024, -0.003215625649318099, 0.07557198405265808, 0.020727548748254776, 0.005025864578783512, -0.07537145912647247, -0.060061465948820114, 0.028297441080212593, 0.029869141057133675, 0.08682557940483093, -0.03418045490980148, 0.015632638707756996, 0.03356447443366051, -0.02039991319179535, -0.0741172656416893, 0.03218350186944008, 0.001260062912479043, 0.06480451673269272, 0.06245157867670059, 0.05904456600546837, -0.0027681365609169006, -0.04995507001876831, 0.27052682638168335, -0.0658140629529953, -0.10584666579961777, -0.13632270693778992, 0.20942670106887817, 0.041670478880405426, -0.03047705814242363, 0.06648743152618408, -0.10619965940713882, -0.03317311406135559, 0.15648934245109558, 0.16800357401371002, -0.03521298989653587, -0.018929891288280487, -0.033611785620450974, -0.01120715495198965, -0.014252842403948307, 0.10471492260694504, 0.09065798670053482, 0.07123656570911407, -0.05300365015864372, -0.014917166903614998, -0.009089901112020016, -0.03861996904015541, -0.08823662251234055, 0.06916138529777527, 0.012456863187253475, -0.0018187548266723752, -0.03410132974386215, 0.08039946109056473, -0.020385922864079475, -0.18736504018306732, 0.04139714688062668, -0.17645558714866638, -0.1860525757074356, -0.004344947636127472, 0.09157232940196991, -0.023676704615354538, 0.04588276892900467, 0.013437160290777683, -0.018525566905736923, 0.10784734785556793, -0.030530964955687523, -0.051773156970739365, -0.10090624541044235, 0.07114719599485397, -0.11454151570796967, 0.2196769267320633, -0.0016108396230265498, 0.08108864724636078, 0.08473183959722519, 0.010907777585089207, -0.13133619725704193, 0.04745633527636528, 0.04814288020133972, -0.07225844264030457, 0.027840454131364822, 0.1781598925590515, -0.04187079519033432, 0.050297483801841736, 0.03287602216005325, -0.11760257184505463, -0.01783020980656147, -0.05067877471446991, -0.01108092162758112, -0.07176728546619415, -0.013469778001308441, -0.026442905887961388, 0.15296931564807892, 0.21604333817958832, -0.017767999321222305, 0.023000109940767288, -0.07161343842744827, -0.006159725598990917, 0.027316533029079437, 0.07915154844522476, -0.01948758214712143, -0.21641646325588226, 0.02447047270834446, -0.02036450430750847, 0.04126884788274765, -0.18668879568576813, -0.08373530209064484, 0.012731152586638927, -0.06196696311235428, -0.051304664462804794, 0.10645531117916107, 0.055528897792100906, 0.035785455256700516, -0.016560319811105728, -0.05982915312051773, -0.011983980424702168, 0.14567740261554718, -0.18619263172149658, -0.03846099227666855 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlnet_gpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "xlnet_gpt2_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/xlnet_gpt2_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# xlnet_gpt2_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# xlnet_gpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# xlnet_gpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 62, 42, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# xlnet_gpt2_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.0842260792851448, 0.14791932702064514, -0.003838211763650179, 0.07087255269289017, 0.12782558798789978, 0.04419068247079849, 0.10726277530193329, 0.12237675487995148, -0.1070202887058258, 0.061321355402469635, 0.08095254749059677, 0.04831302911043167, 0.05437555909156799, 0.1510632038116455, -0.027519090101122856, -0.2189699411392212, 0.0008530659251846373, -0.008020096458494663, -0.0755406841635704, 0.11573118716478348, 0.08931127935647964, -0.09572256356477737, 0.06007152050733566, -0.004888416733592749, -0.1340932548046112, 0.02281108684837818, -0.04336182400584221, -0.05154840275645256, 0.09040789306163788, -0.004672253970056772, 0.09342993795871735, 0.028514482080936432, 0.14547786116600037, -0.2132387012243271, 0.001745798741467297, 0.09592192620038986, 0.03627929091453552, 0.08746365457773209, 0.06019822508096695, -0.010854722931981087, 0.12794780731201172, -0.16142508387565613, 0.09621381014585495, 0.015083836391568184, -0.07456894963979721, -0.1010294109582901, -0.09904464334249496, 0.06817455589771271, 0.09393084049224854, 0.12019268423318863, 0.005491866730153561, 0.11702198535203934, -0.1056448221206665, 0.07944444566965103, 0.1594047099351883, -0.19792555272579193, -0.05381150543689728, 0.07630683481693268, 0.02315116487443447, 0.07054474204778671, -0.08967092633247375, -0.01499507762491703, 0.029993940144777298, 0.024494126439094543, 0.09762032330036163, 0.005905039608478546, -0.09626958519220352, 0.0043119522742927074, -0.12330486625432968, -0.03328027203679085, 0.10011596977710724, 0.02394268475472927, -0.0078188581392169, -0.10213714092969894, -0.052489396184682846, -0.15602536499500275, -0.01762908510863781, -0.010575789026916027, 0.03736509382724762, -0.052594054490327835, -0.051677972078323364, -0.05578171834349632, -0.07181861251592636, -0.05823109671473503, -0.016973525285720825, 0.09971429407596588, 0.042986806482076645, 0.008239643648266792, -0.04761139303445816, 0.12358059734106064, 0.06032714992761612, -0.12524870038032532, 0.0000866023256094195, 0.015627052634954453, -0.08654013276100159, -0.04946627840399742, -0.021397119387984276, -0.022198384627699852, -0.00440349942073226, 0.12724106013774872, -0.07740147411823273, 0.08098713308572769, 0.018724868074059486, -0.012802399694919586, -0.020112313330173492, 0.14238940179347992, -0.0376124270260334, -0.044842708855867386, -0.014831247739493847, 0.11261694878339767, 0.0057427785359323025, -0.010770522058010101, -0.07091205567121506, -0.0027041290886700153, 0.06438995897769928, 0.06361711025238037, -0.04389602318406105, 0.04126406088471413, -0.02039637416601181, -0.01587722823023796, 0.042891982942819595, -0.1389589011669159, 0.03976720944046974, 0.015016546472907066, -0.09776191413402557, -0.01161231566220522, 0.01450431626290083, -0.004199817311018705, -0.05055948346853256, 0.11195743083953857, -0.07600138336420059, 0.005187348462641239, -0.08520600944757462, -0.06929941475391388, 0.016498779878020287, -0.09889719635248184, -0.04161324352025986, -0.03364916890859604, -0.19278375804424286, -0.06317798048257828, 0.023098725825548172, -0.05571784824132919, -0.04251009598374367, -0.06387574970722198, -0.07934818416833878, 0.018485670909285545, -0.006147436797618866, 0.1769498586654663, -0.06610491871833801, 0.06459839642047882, -0.022049549967050552, 0.023658256977796555, 0.0640583261847496, 0.0427505187690258, -0.06097335368394852, -0.00046941294567659497, -0.09389972686767578, 0.09181735664606094, -0.08493636548519135, 0.00016284009325318038, -0.10869444906711578, -0.093756765127182, 0.014040909707546234, -0.011830298230051994, 0.046524226665496826, 0.13700225949287415, -0.1930057853460312, -0.030213218182325363, 0.13016003370285034, -0.05473167076706886, -0.00616664020344615, 0.05705142766237259, -0.052700333297252655, -0.008226984180510044, 0.060288455337285995, 0.12568528950214386, 0.0829213410615921, -0.12497824430465698, -0.01021923404186964, 0.007481093518435955, 0.03520860895514488, 0.009607290849089622, 0.02054346725344658, 0.00505858613178134, 0.05929084122180939, 0.008576416410505772, -0.05501272901892662, 0.009071205742657185, -0.09196295589208603, -0.06896843761205673, -0.038822710514068604, -0.07640071958303452, 0.020351583138108253, 0.02461962029337883, 0.03415779769420624, -0.05616569519042969, -0.11381889134645462, 0.0819941908121109, 0.13093848526477814, -0.056283995509147644, 0.007339976262301207, -0.07787289470434189, -0.042500920593738556, 0.008593504317104816, -0.020679421722888947, -0.19107744097709656, -0.11994887888431549, 0.02410884201526642, -0.06087664142251015, 0.041993893682956696, -0.012968281283974648, 0.06512390822172165, 0.041580818593502045, -0.04092632234096527, -0.0215108972042799, -0.07192986458539963, -0.0033993027172982693, -0.08755754679441452, -0.2000628411769867, -0.03064822219312191, -0.015381296165287495, 0.17693781852722168, -0.2341029942035675, 0.022729603573679924, -0.0070191798731684685, 0.14477187395095825, 0.02169438637793064, -0.059040822088718414, 0.018390603363513947, 0.05020909756422043, 0.0022609580773860216, -0.0955202504992485, 0.023494739085435867, -0.011558893136680126, -0.07875315099954605, -0.017229704186320305, -0.131593719124794, -0.02233363874256611, 0.07453744858503342, 0.10155533254146576, -0.10452786833047867, 0.014442682266235352, -0.07163368910551071, -0.04781905189156532, -0.0952233225107193, 0.02778318151831627, 0.2066008597612381, 0.03186216577887535, 0.12816141545772552, -0.04844803363084793, -0.08490303158760071, -0.005388368386775255, 0.02307703346014023, 0.005614993162453175, 0.10208582878112793, 0.07765839248895645, -0.0422782301902771, 0.07674524933099747, 0.010948077775537968, -0.04890883341431618, 0.14316213130950928, -0.04643741995096207, -0.08668067306280136, -0.0038867995608597994, -0.015229461714625359, -0.02533453144133091, 0.1035487949848175, -0.046713218092918396, -0.0005009045707993209, 0.028590457513928413, 0.024262776598334312, 0.03977318853139877, -0.1848936378955841, -0.0017884395783767104, 0.015415576286613941, -0.06781504303216934, -0.03100435808300972, -0.025546014308929443, 0.05025271698832512, 0.09929285943508148, 0.0003490403469186276, -0.02781330607831478, 0.013417011126875877, -0.009360058233141899, -0.07439561933279037, 0.17438748478889465, -0.11237886548042297, -0.17121200263500214, -0.09844991564750671, 0.0459873303771019, -0.035191796720027924, -0.04988737031817436, -0.0029536026995629072, -0.11379226297140121, -0.06622374802827835, -0.1178026869893074, -0.019071726128458977, -0.0015114162815734744, -0.009282737970352173, 0.04155914857983589, 0.020215697586536407, 0.04903138801455498, -0.13430559635162354, 0.016370516270399094, -0.03721217066049576, -0.08080600947141647, 0.016604090109467506, 0.06620695441961288, 0.06519477814435959, 0.1524873971939087, -0.015147092752158642, 0.03083779290318489, -0.016909807920455933, 0.185947984457016, -0.10129974037408829, 0.015196261927485466, 0.10497689992189407, 0.022023223340511322, 0.04246903583407402, 0.11076987534761429, 0.03417043387889862, -0.07393021881580353, 0.027076084166765213, 0.08926056325435638, -0.0237130019813776, -0.26426902413368225, -0.06341452151536942, -0.01736137457191944, -0.07419631630182266, 0.08781763911247253, 0.060150086879730225, 0.005674086045473814, 0.009325971826910973, -0.010391482152044773, -0.00680120661854744, -0.0077642155811190605, 0.05479462072253227, 0.08227893710136414, 0.04691142588853836, 0.08564700186252594, -0.03491123393177986, -0.023631490767002106, 0.06531393527984619, 0.007174077909439802, 0.22143130004405975, -0.0606536939740181, 0.0873551145195961, 0.014862319454550743, 0.10086184740066528, -0.012772497721016407, 0.022732097655534744, 0.02799515798687935, -0.008163457736372948, 0.018631165847182274, -0.04761545732617378, -0.0036640155594795942, 0.007892824709415436, -0.010145503096282482, -0.006381910294294357, -0.058741532266139984, 0.04064213111996651, 0.011632468551397324, 0.28281742334365845, 0.037545256316661835, -0.2899005115032196, -0.05206070840358734, -0.017743635922670364, -0.03110656514763832, -0.054272253066301346, -0.008169732056558132, 0.1114618182182312, -0.11622962355613708, 0.06614577770233154, -0.03968251124024391, 0.08469158411026001, -0.07765534520149231, -0.005509085021913052, 0.06006757542490959, 0.13519717752933502, -0.01553169172257185, 0.06870705634355545, -0.20990176498889923, 0.2130495011806488, 0.010029951110482216, 0.12334086745977402, -0.07014620304107666, 0.023133713752031326, 0.012432089075446129, 0.04504654183983803, 0.07546145468950272, -0.005554979667067528, -0.030171416699886322, -0.14367131888866425, -0.1267090141773224, 0.024774381890892982, 0.11448805779218674, -0.01177933532744646, 0.08084101229906082, -0.016713818535208702, -0.01597345434129238, 0.03734465688467026, -0.16808266937732697, -0.1540904939174652, -0.1256207376718521, 0.031209804117679596, 0.04167146235704422, -0.03550048917531967, -0.058816730976104736, -0.11111722141504288, -0.017884256318211555, 0.1793043613433838, -0.006048998329788446, -0.062394727021455765, -0.14376312494277954, 0.0642637312412262, 0.14703601598739624, -0.038149215281009674, 0.009350406005978584, 0.039543960243463516, 0.12816940248012543, 0.04529762268066406, -0.07562248408794403, 0.027835320681333542, -0.05909808352589607, -0.18553605675697327, -0.053934965282678604, 0.1371767669916153, 0.08887230604887009, 0.047979686409235, -0.0194404199719429, 0.020280392840504646, 0.00556588638573885, -0.08254807442426682, 0.005197147373110056, 0.09148655831813812, 0.08121806383132935, 0.06638544052839279, -0.06109458580613136, 0.0018817895324900746, -0.04464532807469368, -0.030514108017086983, 0.10176390409469604, 0.1762005090713501, -0.0779089629650116, 0.09972994029521942, 0.0556180402636528, -0.07484713196754456, -0.1736064851284027, 0.0570073239505291, 0.13009881973266602, 0.018594399094581604, 0.025480154901742935, -0.20737485587596893, 0.1149928867816925, 0.12798917293548584, -0.01208472903817892, 0.06120790168642998, -0.35916125774383545, -0.12418342381715775, 0.04682810232043266, 0.07732287049293518, -0.00631409976631403, -0.10489196330308914, -0.028737328946590424, -0.010689783841371536, -0.1364661604166031, 0.1354367882013321, -0.06374351680278778, 0.10445871949195862, -0.010663589462637901, 0.11276616156101227, 0.025403648614883423, -0.04738406091928482, 0.12126412242650986, 0.07234065979719162, 0.06757853925228119, -0.04532850533723831, 0.01852203905582428, 0.051921091973781586, -0.07101558148860931, 0.05429486185312271, -0.05208713188767433, 0.07104271650314331, -0.13317488133907318, -0.018448252230882645, -0.0811031237244606, 0.0426730141043663, -0.04559429734945297, -0.04049260541796684, -0.04035166651010513, 0.05007408186793327, 0.06625845283269882, -0.041821256279945374, 0.07050836831331253, 0.0051851118914783, 0.0779472067952156, 0.08213754743337631, 0.08917396515607834, -0.02996188960969448, -0.09923519194126129, -0.011024117469787598, -0.008895727805793285, 0.035999614745378494, -0.12752018868923187, 0.019790608435869217, 0.13584181666374207, 0.05243538320064545, 0.12653395533561707, 0.02976738102734089, -0.040355533361434937, -0.01883685030043125, 0.036844298243522644, -0.11867654323577881, -0.1078525260090828, 0.020628102123737335, -0.0612490177154541, -0.11109733581542969, 0.008085140958428383, 0.10511195659637451, -0.04011138528585434, -0.014737802557647228, -0.0006129404646344483, 0.03207004815340042, 0.0071026538498699665, 0.21222124993801117, 0.026154251769185066, 0.06614911556243896, -0.10992815345525742, 0.1210251897573471, 0.06830360740423203, -0.09462545067071915, 0.04687250778079033, 0.12664794921875, -0.09565884619951248, -0.014093359932303429, 0.09567232429981232, 0.1415673792362213, -0.03636258468031883, -0.039585717022418976, -0.08873782306909561, -0.09775020182132721, 0.07869400084018707, 0.11949250847101212, 0.029908057302236557, 0.011939235031604767, -0.04832211136817932, 0.011949810199439526, -0.14917826652526855, 0.06494779884815216, 0.07885609567165375, 0.056871652603149414, -0.09288208186626434, 0.13849306106567383, 0.025233617052435875, 0.03547291457653046, -0.014704158529639244, 0.01386483944952488, -0.06575290858745575, -0.0068253022618591785, -0.08769670873880386, -0.002230249810963869, -0.006389258895069361, -0.003512119874358177, -0.017073798924684525, -0.042994868010282516, -0.040398553013801575, 0.04477108642458916, -0.0715940073132515, -0.06014883145689964, -0.009631596505641937, 0.046327169984579086, -0.1299377977848053, 0.004457807634025812, 0.001250788918696344, -0.09609942138195038, 0.07103375345468521, 0.046919748187065125, 0.0007569625740870833, 0.030291376635432243, -0.13462544977664948, -0.023498328402638435, 0.03320224955677986, 0.02316633053123951, 0.06857249140739441, -0.06533045321702957, 0.0010502865770831704, -0.021101580932736397, 0.03210357576608658, 0.01882457174360752, 0.03681913763284683, -0.11382932960987091, 0.007708938326686621, -0.06793539226055145, -0.0300579946488142, -0.07317093014717102, 0.04651094600558281, 0.11289206147193909, 0.03984932601451874, 0.16747018694877625, -0.0788576677441597, 0.03540762513875961, -0.18577513098716736, -0.028504423797130585, 0.0004380055470392108, -0.03358118236064911, -0.05297841876745224, -0.01431939098984003, 0.10886862128973007, -0.05055752396583557, 0.12665492296218872, 0.011299651116132736, 0.06223205476999283, 0.03899076581001282, -0.034078408032655716, -0.06081965193152428, 0.014833031222224236, 0.14156252145767212, 0.06203955411911011, -0.020743144676089287, 0.1120021641254425, -0.02100484073162079, 0.05032080411911011, 0.04682515934109688, 0.21813519299030304, 0.12421796470880508, 0.02156199887394905, 0.07427718490362167, 0.056449878960847855, -0.14003103971481323, -0.11668732017278671, 0.11801300942897797, -0.09049077332019806, 0.12514106929302216, -0.06378762423992157, 0.19495771825313568, 0.05995943769812584, -0.15025568008422852, 0.0520559661090374, -0.028217192739248276, -0.10856331884860992, -0.09868582338094711, -0.021792689338326454, -0.08036770671606064, -0.10442818701267242, 0.03553010895848274, -0.11482789367437363, 0.0617271363735199, 0.0948152169585228, 0.03183376044034958, 0.02139916643500328, 0.11608804017305374, 0.0014738942263647914, -0.00481176795437932, 0.06572592258453369, 0.02227836661040783, 0.0012744407868012786, -0.07573942095041275, -0.06188216060400009, 0.03349713981151581, 0.024642247706651688, 0.08771857619285583, -0.03378862142562866, 0.0023572605568915606, 0.03832501173019409, -0.013914918527007103, -0.06828280538320541, 0.03569398075342178, 0.003234811360016465, 0.0609748475253582, 0.06051122024655342, 0.05552518740296364, 0.0010208021849393845, -0.04704644903540611, 0.27495065331459045, -0.06210298091173172, -0.10895437747240067, -0.12783436477184296, 0.21700042486190796, 0.05012736842036247, -0.020430099219083786, 0.0644390657544136, -0.10658726841211319, -0.03970622271299362, 0.1523580104112625, 0.16509737074375153, -0.05188252776861191, -0.02757880464196205, -0.030644072219729424, -0.012817492708563805, -0.027458686381578445, 0.11531831324100494, 0.10462403297424316, 0.06500788778066635, -0.047009240835905075, -0.019051216542720795, -0.010588038712739944, -0.03549923375248909, -0.07162435352802277, 0.07226022332906723, 0.01579340361058712, -0.00324674928560853, -0.033737912774086, 0.07445258647203445, -0.017891010269522667, -0.19488191604614258, 0.04578906297683716, -0.16568782925605774, -0.18642182648181915, -0.004395224153995514, 0.09612292051315308, -0.028559014201164246, 0.0438244454562664, 0.010989924892783165, -0.01713116094470024, 0.10427523404359818, -0.03644280135631561, -0.04405871406197548, -0.09973684698343277, 0.06121727079153061, -0.11560825258493423, 0.2330513745546341, -0.004593443125486374, 0.07505681365728378, 0.09003445506095886, 0.010549834929406643, -0.13041634857654572, 0.06175490468740463, 0.04451107978820801, -0.08034300804138184, 0.030425990000367165, 0.1674029678106308, -0.04577399417757988, 0.04483749717473984, 0.034906983375549316, -0.1124315932393074, -0.010010522790253162, -0.06922396272420883, -0.02354525215923786, -0.07155455648899078, -0.018642231822013855, -0.029596664011478424, 0.14805543422698975, 0.21176478266716003, -0.01115403976291418, 0.028750818222761154, -0.07271046191453934, 0.000661442696582526, 0.023275114595890045, 0.09738772362470627, -0.018575305119156837, -0.21337351202964783, 0.02738877572119236, 0.0028736479580402374, 0.034200362861156464, -0.1895834356546402, -0.07255404442548752, 0.0084415627643466, -0.06473745405673981, -0.04589402303099632, 0.10770078748464584, 0.055546488612890244, 0.04213547706604004, -0.020454969257116318, -0.07526355236768723, -0.011474469676613808, 0.14402097463607788, -0.18418511748313904, -0.04104512557387352 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlnet_gpt2_summarization_xsum This model is a fine-tuned version of [](https://huggingface.co/) on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["xsum"], "model-index": [{"name": "xlnet_gpt2_summarization_xsum", "results": []}]}
text2text-generation
Ayham/xlnet_gpt2_summarization_xsum
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:xsum", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us
# xlnet_gpt2_summarization_xsum This model is a fine-tuned version of [](URL on the xsum dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# xlnet_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us \n", "# xlnet_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 59, 36, 6, 12, 8, 3, 118, 4, 36 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-xsum #autotrain_compatible #endpoints_compatible #region-us \n# xlnet_gpt2_summarization_xsum\n\nThis model is a fine-tuned version of [](URL on the xsum dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.08765918016433716, 0.14003905653953552, -0.002609168877825141, 0.06081331893801689, 0.1491285264492035, 0.0444101020693779, 0.09916073083877563, 0.11639870703220367, -0.09115397185087204, 0.07906771451234818, 0.0788351446390152, 0.0720011442899704, 0.06217549741268158, 0.1511375606060028, -0.03849629685282707, -0.24287691712379456, 0.020680122077465057, -0.024272721260786057, -0.086654894053936, 0.09928548336029053, 0.0882846936583519, -0.11338154971599579, 0.07187087088823318, -0.0041213505901396275, -0.16435547173023224, 0.010702071711421013, -0.039500534534454346, -0.04857715964317322, 0.09727793186903, 0.006650546099990606, 0.09496951103210449, 0.019995441660284996, 0.13236023485660553, -0.23005421459674835, 0.003449587384238839, 0.09541677683591843, 0.037994395941495895, 0.08736615628004074, 0.0585828572511673, 0.004604953806847334, 0.13460227847099304, -0.14173269271850586, 0.0991920754313469, 0.021236762404441833, -0.07570582628250122, -0.11978157609701157, -0.09268016368150711, 0.026699673384428024, 0.07880078256130219, 0.09482655674219131, 0.005713086575269699, 0.12245946377515793, -0.09665403515100479, 0.07360325753688812, 0.16688740253448486, -0.23016761243343353, -0.06005178019404411, 0.035256799310445786, 0.06734873354434967, 0.07143832743167877, -0.1053759753704071, -0.00856660958379507, 0.02211691439151764, 0.024784201756119728, 0.08875234425067902, -0.0030630335677415133, -0.10240087658166885, 0.010720531456172466, -0.12424582988023758, -0.01215536892414093, 0.09679271280765533, 0.031002609059214592, -0.02440241165459156, -0.10961844772100449, -0.058578670024871826, -0.11730197817087173, -0.014213736169040203, -0.029829241335392, 0.04571548104286194, -0.0387469045817852, -0.06912633031606674, -0.04567226022481918, -0.06477808207273483, -0.059018488973379135, -0.021252119913697243, 0.14219774305820465, 0.03172314167022705, 0.014877565205097198, -0.03640018776059151, 0.0958036556839943, 0.044522348791360855, -0.11022654175758362, -0.0024506039917469025, -0.005953056272119284, -0.11251553148031235, -0.04737416282296181, -0.054896168410778046, -0.03017987497150898, 0.01560573372989893, 0.143434077501297, -0.06461135298013687, 0.09397387504577637, 0.02513529546558857, -0.014199280180037022, 0.0028434074483811855, 0.1493178904056549, -0.05171440169215202, -0.056955259293317795, -0.018122173845767975, 0.08526269346475601, 0.014995642937719822, -0.023228203877806664, -0.063352070748806, -0.0043675838969647884, 0.07315423339605331, 0.0580395832657814, -0.04252985864877701, 0.028608839958906174, -0.03907307609915733, -0.026464121416211128, 0.0012036588741466403, -0.1239529550075531, 0.05349406972527504, 0.002723902463912964, -0.08515587449073792, 0.001935996231622994, 0.0010710540227591991, 0.012490401975810528, -0.0357632078230381, 0.118623286485672, -0.08231138437986374, -0.005430466495454311, -0.09187807142734528, -0.07965568453073502, 0.021747643128037453, -0.13204364478588104, -0.014699537307024002, -0.03591595217585564, -0.1667473316192627, -0.05322067812085152, 0.06466732174158096, -0.056965816766023636, -0.02170489728450775, -0.050105709582567215, -0.044086236506700516, 0.038983892649412155, -0.012798244133591652, 0.16899506747722626, -0.05964620038866997, 0.0674692913889885, -0.03375405818223953, 0.03692201152443886, 0.008257078006863594, 0.04794960841536522, -0.0710662305355072, 0.00528927193954587, -0.1071682944893837, 0.08389413356781006, -0.0672188326716423, 0.009457550011575222, -0.11359421163797379, -0.0836060643196106, -0.006568167824298143, -0.005701298359781504, 0.07526031881570816, 0.11645197868347168, -0.2006988525390625, -0.03153057023882866, 0.11115545779466629, -0.07038365304470062, -0.057328831404447556, 0.05975693464279175, -0.049684006720781326, 0.0061878906562924385, 0.048197511583566666, 0.15636175870895386, 0.04424667730927467, -0.12056245654821396, -0.03147754445672035, 0.0028672886546701193, 0.043730251491069794, 0.01659155637025833, 0.044213782995939255, -0.00532690342515707, 0.06443548202514648, 0.00872261356562376, -0.03410816192626953, 0.005649720784276724, -0.07828088849782944, -0.07646088302135468, -0.03649131953716278, -0.07179151475429535, -0.0022680978290736675, 0.02525758184492588, 0.02092263102531433, -0.05826131999492645, -0.10818985849618912, 0.10604368895292282, 0.1243346706032753, -0.07633839547634125, 0.022165926173329353, -0.06839974969625473, -0.01569512113928795, -0.003102074610069394, -0.020538436248898506, -0.20771941542625427, -0.11435290426015854, 0.02770807035267353, -0.061294835060834885, 0.031087294220924377, -0.016566848382353783, 0.05772752687335014, 0.04954679310321808, -0.029454201459884644, -0.010495380498468876, -0.06903219223022461, -0.0019621418323367834, -0.09517250955104828, -0.21974971890449524, -0.04672961309552193, -0.022667182609438896, 0.1880393773317337, -0.20897360146045685, -0.006002964451909065, -0.004863252863287926, 0.1399405151605606, 0.028222858905792236, -0.06971180438995361, -0.0020639155991375446, 0.04736226052045822, -0.011949395760893822, -0.09756768494844437, 0.03720727190375328, 0.012154029682278633, -0.09274239093065262, -0.038241758942604065, -0.15155522525310516, -0.003946118988096714, 0.09820354729890823, 0.054498013108968735, -0.07955339550971985, -0.03440301865339279, -0.06630387157201767, -0.03857949748635292, -0.07169921696186066, 0.0268523171544075, 0.18413661420345306, 0.020209593698382378, 0.1101875901222229, -0.06149846315383911, -0.07226503640413284, 0.008749273605644703, 0.03418179601430893, -0.0061327191069722176, 0.08934931457042694, 0.12855109572410583, -0.08207429200410843, 0.08296908438205719, 0.0676274225115776, -0.03276138752698898, 0.1405762881040573, -0.04053093120455742, -0.07991443574428558, -0.0073794955387711525, -0.007349285762757063, -0.022261036559939384, 0.12029464542865753, -0.08759105205535889, 0.01555553637444973, 0.027658233419060707, 0.03395020216703415, 0.04788805916905403, -0.18051473796367645, 0.011003095656633377, 0.016832059249281883, -0.03954780474305153, -0.04724571853876114, -0.022992322221398354, 0.03190295770764351, 0.089338518679142, 0.024453353136777878, 0.004245928954333067, 0.011248617433011532, -0.01101293321698904, -0.08094829320907593, 0.18748804926872253, -0.13929583132266998, -0.1639157384634018, -0.07935058325529099, 0.03221182897686958, -0.04696769639849663, -0.03601771965622902, 0.003156352322548628, -0.10294100642204285, -0.06435078382492065, -0.09114226698875427, -0.024783503264188766, -0.03489502891898155, 0.011099123395979404, 0.04464729130268097, 0.014659896492958069, 0.04541100561618805, -0.13450413942337036, 0.013599525205790997, -0.056521154940128326, -0.06405339390039444, 0.00972111988812685, 0.08065538108348846, 0.08809079974889755, 0.13025152683258057, -0.01661412976682186, 0.022833412513136864, -0.026427410542964935, 0.21035006642341614, -0.08982151001691818, 0.0021613663993775845, 0.10931391268968582, -0.004408280830830336, 0.044882941991090775, 0.11850626766681671, 0.03320879861712456, -0.09511887282133102, 0.030613135546445847, 0.07305806875228882, -0.0201723575592041, -0.24722374975681305, -0.05144402012228966, -0.039689768105745316, -0.09658906608819962, 0.09548614919185638, 0.04533635452389717, -0.04352483153343201, 0.037182144820690155, -0.00416989903897047, 0.02218102663755417, -0.019002625718712807, 0.05784066021442413, 0.07796109467744827, 0.054514043033123016, 0.1020902767777443, -0.022616323083639145, -0.011989980936050415, 0.0729200541973114, 0.002955242060124874, 0.25874918699264526, -0.034264687448740005, 0.0898766815662384, 0.03107338771224022, 0.10527736693620682, -0.020234523341059685, 0.05350949615240097, 0.015426735393702984, -0.007177598774433136, -0.0050639561377465725, -0.05934065952897072, -0.034867651760578156, 0.022964730858802795, -0.012401865795254707, 0.004714444745332003, -0.08788024634122849, 0.04899982735514641, 0.032065656036138535, 0.24037902057170868, 0.02210080996155739, -0.2887972593307495, -0.05420173332095146, -0.004437593277543783, -0.02991345338523388, -0.06716859340667725, -0.006143426056951284, 0.1188955008983612, -0.11287686228752136, 0.07356172800064087, -0.06821954995393753, 0.08532929420471191, -0.04541751742362976, -0.003750626230612397, 0.07488180696964264, 0.16038382053375244, -0.010038643144071102, 0.05847376585006714, -0.20242303609848022, 0.21462887525558472, 0.02090279571712017, 0.1228363886475563, -0.07827901095151901, 0.04516326263546944, 0.012723620980978012, 0.011345274746418, 0.07535766810178757, -0.0012763876002281904, -0.10911241173744202, -0.13684186339378357, -0.07990118861198425, 0.06315591186285019, 0.1523313671350479, -0.008629398420453072, 0.08983580023050308, -0.043350912630558014, 0.011092837899923325, 0.047368187457323074, -0.09151463955640793, -0.15726380050182343, -0.1517345905303955, 0.026095209643244743, 0.020420413464307785, -0.03147006407380104, -0.04851881414651871, -0.09454578906297684, -0.03125886246562004, 0.1689477562904358, -0.016497189179062843, -0.0468527115881443, -0.1582147777080536, 0.0769384503364563, 0.16514162719249725, -0.05195970460772514, 0.029912376776337624, 0.015477584674954414, 0.12057986855506897, 0.03359160199761391, -0.091032013297081, 0.060061559081077576, -0.07549397647380829, -0.16146783530712128, -0.06133076176047325, 0.10300185531377792, 0.05881941691040993, 0.039935894310474396, -0.01500692404806614, 0.03961201757192612, -0.009857868775725365, -0.09694066643714905, 0.024347536265850067, 0.09687399864196777, 0.06723093241453171, 0.06649234145879745, -0.08838994801044464, 0.031048189848661423, -0.013157133013010025, -0.026136303320527077, 0.12702301144599915, 0.19516268372535706, -0.0753290057182312, 0.1008882150053978, 0.08017919957637787, -0.08809604495763779, -0.18572430312633514, 0.08402294665575027, 0.11062353849411011, 0.02393685281276703, 0.04727746918797493, -0.22746630012989044, 0.13184691965579987, 0.1256885528564453, -0.00600254163146019, 0.057737529277801514, -0.303998738527298, -0.12697456777095795, 0.03712243586778641, 0.10325610637664795, 0.017671547830104828, -0.12067773193120956, -0.020104430615901947, -0.039926689118146896, -0.12042953819036484, 0.15176431834697723, -0.07454909384250641, 0.11143285036087036, 0.004959185142070055, 0.07690785080194473, 0.022990701720118523, -0.04449581354856491, 0.12224458158016205, 0.025356529280543327, 0.06860847026109695, -0.03443311154842377, 0.013145838864147663, 0.004148636944591999, -0.059456717222929, 0.024721089750528336, -0.08051519095897675, 0.06144552677869797, -0.10864967852830887, -0.01390352938324213, -0.05752415210008621, 0.050063733011484146, -0.04231167957186699, -0.05902398005127907, -0.04069528728723526, 0.050200674682855606, 0.08176374435424805, -0.03280523419380188, 0.0630510002374649, 0.004524717107415199, 0.08165336400270462, 0.047497332096099854, 0.10172444581985474, -0.07432081550359726, -0.06042443960905075, 0.008498383685946465, -0.0050816782750189304, 0.04734821617603302, -0.11389512568712234, 0.02840575762093067, 0.14095664024353027, 0.04445667564868927, 0.12721949815750122, 0.056260157376527786, -0.040662165731191635, -0.00021497803390957415, 0.05095425248146057, -0.11466242372989655, -0.11881883442401886, 0.015248777344822884, -0.04650042951107025, -0.12243450433015823, 0.017605174332857132, 0.10930874198675156, -0.026986384764313698, -0.00743288965895772, -0.014355015009641647, 0.038041599094867706, 0.007444364950060844, 0.18893256783485413, 0.010673005133867264, 0.05361491069197655, -0.10867312550544739, 0.13733208179473877, 0.0630752220749855, -0.11227277666330338, 0.06314648687839508, 0.09656757116317749, -0.08699265122413635, 0.00014477840159088373, 0.07126688212156296, 0.12148973345756531, -0.038218945264816284, -0.043799322098493576, -0.08888150006532669, -0.09813588112592697, 0.055981673300266266, 0.11119478195905685, 0.028775185346603394, 0.0056219580583274364, -0.043591249734163284, 0.03399575874209404, -0.15222027897834778, 0.06497615575790405, 0.05061672255396843, 0.06154683232307434, -0.11265677213668823, 0.1185661181807518, 0.020534707233309746, 0.011194274760782719, -0.014119251631200314, 0.01877937838435173, -0.09618285298347473, -0.02380293235182762, -0.07969441264867783, -0.012437324970960617, -0.035870473831892014, -0.0029185842722654343, -0.005659126676619053, -0.037859462201595306, -0.05695688724517822, 0.034123845398426056, -0.0742272213101387, -0.060453541576862335, -0.006144983693957329, 0.03659161552786827, -0.13741633296012878, 0.00821745302528143, 0.011012077331542969, -0.09900737553834915, 0.08135774731636047, 0.052105870097875595, 0.017963832244277, 0.03489162027835846, -0.1599450558423996, -0.04357938840985298, 0.02805016003549099, 0.024482427164912224, 0.07066012173891068, -0.08966781198978424, -0.007319137919694185, -0.0049512325786054134, 0.04585381597280502, 0.008842349983751774, 0.0529639795422554, -0.11253153532743454, 0.00018368198652751744, -0.07430364936590195, -0.04729018360376358, -0.060835130512714386, 0.047021813690662384, 0.11055044829845428, 0.039972398430109024, 0.16529542207717896, -0.06360886991024017, 0.031860802322626114, -0.19071413576602936, -0.02242916077375412, -0.006817561574280262, -0.03228386119008064, -0.08644729107618332, -0.020802097395062447, 0.09542720019817352, -0.05048627033829689, 0.13554619252681732, 0.010983222164213657, 0.07020873576402664, 0.036247022449970245, -0.030237436294555664, -0.05356164276599884, 0.014224919490516186, 0.19028158485889435, 0.0804109200835228, -0.017355868592858315, 0.09819474071264267, 0.0026349839754402637, 0.05996256321668625, 0.06210631504654884, 0.23763389885425568, 0.15340138971805573, 0.020425714552402496, 0.08717820793390274, 0.05496583878993988, -0.13105309009552002, -0.12243830412626266, 0.1558721512556076, -0.06819045543670654, 0.12478171288967133, -0.05429026111960411, 0.18701577186584473, 0.05920223519206047, -0.17075258493423462, 0.03504261001944542, -0.04254620149731636, -0.10059250146150589, -0.12854835391044617, -0.0032463697716593742, -0.08368775993585587, -0.1181723028421402, 0.021023133769631386, -0.1295609474182129, 0.05893620848655701, 0.08430323749780655, 0.018881643190979958, 0.03342593461275101, 0.10233709961175919, -0.02739308401942253, 0.0009404915035702288, 0.048966772854328156, 0.03580048680305481, 0.00015024877211544663, -0.043863434344530106, -0.08662204444408417, 0.03073956072330475, 0.013660344295203686, 0.097053661942482, -0.045717306435108185, -0.013894363306462765, 0.05407941713929176, -0.01147120539098978, -0.06240148842334747, 0.021453315392136574, 0.004484960343688726, 0.044952452182769775, 0.043455734848976135, 0.04698308929800987, -0.004974175710231066, -0.04782574996352196, 0.2695596218109131, -0.0650983527302742, -0.07395274937152863, -0.13413198292255402, 0.1983049213886261, 0.036754071712493896, -0.01391415111720562, 0.06809598952531815, -0.10513085126876831, -0.044407833367586136, 0.16375699639320374, 0.15775653719902039, -0.08650444447994232, -0.03128686547279358, -0.007496669422835112, -0.010333786718547344, -0.0526459664106369, 0.14390072226524353, 0.09804626554250717, 0.07874678820371628, -0.04952777177095413, -0.019067654386162758, -0.01839998923242092, -0.0438808910548687, -0.05671992152929306, 0.05339639633893967, 0.019559437409043312, -0.008909805677831173, -0.02805870585143566, 0.06976623088121414, -0.010673854500055313, -0.18519558012485504, 0.03712770715355873, -0.18051598966121674, -0.17983286082744598, -0.020086336880922318, 0.09703559428453445, -0.028249826282262802, 0.0481751412153244, -0.00003363321957294829, -0.01867455430328846, 0.10025721043348312, -0.01001804694533348, -0.047596707940101624, -0.09813878685235977, 0.07637841254472733, -0.07166699320077896, 0.22404448688030243, -0.012270215898752213, 0.06354987621307373, 0.10214102268218994, 0.05131029710173607, -0.1282368004322052, 0.026645900681614876, 0.05928497388958931, -0.10810792446136475, 0.040055714547634125, 0.13725650310516357, -0.061023786664009094, 0.07473690062761307, 0.04899153485894203, -0.10068642348051071, 0.005739255342632532, -0.0687892809510231, -0.03927624970674515, -0.05456497520208359, -0.020482612773776054, -0.0655069574713707, 0.1659071445465088, 0.20684698224067688, -0.015331140719354153, 0.020282968878746033, -0.07573864609003067, 0.031021330505609512, 0.025406090542674065, 0.11672735959291458, -0.03634873405098915, -0.23454329371452332, 0.028078481554985046, 0.028376774862408638, 0.015442120842635632, -0.18772642314434052, -0.07317651063203812, 0.028684284538030624, -0.0537363663315773, -0.0645093098282814, 0.10643390566110611, 0.06078200414776802, 0.024561380967497826, -0.041140347719192505, -0.10325625538825989, -0.022816333919763565, 0.1444953978061676, -0.14773593842983246, -0.04232171177864075 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlnet_gpt_xsum This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "xlnet_gpt_xsum", "results": []}]}
text2text-generation
Ayham/xlnet_gpt_xsum
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
# xlnet_gpt_xsum This model is a fine-tuned version of [](URL on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# xlnet_gpt_xsum\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n", "# xlnet_gpt_xsum\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 53, 32, 6, 12, 8, 3, 118, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n# xlnet_gpt_xsum\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.08017653226852417, 0.13802719116210938, -0.0028217590879648924, 0.059507932513952255, 0.14498843252658844, 0.04247729480266571, 0.10195133090019226, 0.11627209931612015, -0.08257497847080231, 0.07818805426359177, 0.08225485682487488, 0.06448142230510712, 0.05959369242191315, 0.11474838107824326, -0.027567705139517784, -0.24522925913333893, 0.021453959867358208, 0.004418298602104187, -0.07381978631019592, 0.09280498325824738, 0.09561605751514435, -0.09351875633001328, 0.06660976260900497, 0.008467912673950195, -0.16018880903720856, 0.02669515274465084, -0.056043677031993866, -0.0667872205376625, 0.08436542004346848, 0.02323644421994686, 0.08988875150680542, 0.011207823641598225, 0.09546276926994324, -0.22955943644046783, 0.005339103285223246, 0.07694027572870255, 0.02674221247434616, 0.0865878313779831, 0.07730402052402496, 0.0008964607841335237, 0.14260421693325043, -0.14693592488765717, 0.08697296679019928, 0.029342029243707657, -0.04625938460230827, -0.12190531194210052, -0.06946200877428055, 0.07900147885084152, 0.0666186660528183, 0.09674476832151413, 0.00991801731288433, 0.13146916031837463, -0.09699446707963943, 0.07613201439380646, 0.19485357403755188, -0.2520481050014496, -0.05835571885108948, 0.034408532083034515, 0.07560787349939346, 0.08977137506008148, -0.09241911768913269, -0.011718638241291046, 0.04270561784505844, 0.03694240003824234, 0.07744079828262329, 0.0025731970090419054, -0.06388276815414429, -0.00040870255907066166, -0.1387234777212143, -0.014405724592506886, 0.1564003825187683, 0.05013453960418701, -0.03150706738233566, -0.11082367599010468, -0.05082050338387489, -0.10602357983589172, -0.016782894730567932, -0.03603225573897362, 0.028780722990632057, -0.029632825404405594, -0.05867304652929306, -0.07232724130153656, -0.06654838472604752, -0.05682538449764252, -0.01164033729583025, 0.17857521772384644, 0.05987812206149101, -0.000358612829586491, -0.030809342861175537, 0.09352874755859375, 0.06582103669643402, -0.10574381798505783, -0.008313313126564026, -0.013785294257104397, -0.062369976192712784, -0.04254336655139923, -0.057943232357501984, -0.03634250909090042, 0.034165870398283005, 0.12432791292667389, -0.0749579519033432, 0.09072752296924591, 0.02400015853345394, 0.012660603038966656, 0.002348770620301366, 0.12842385470867157, -0.0528591051697731, -0.023577244952321053, -0.007327173836529255, 0.0918797180056572, 0.01944603957235813, -0.021288851276040077, -0.09235275536775589, -0.036816105246543884, 0.09376080334186554, 0.06683853268623352, -0.03736944496631622, 0.027176400646567345, -0.04753037914633751, -0.03909473866224289, 0.002465640427544713, -0.12319900095462799, 0.04425213485956192, 0.004343064036220312, -0.08191809803247452, 0.02568008378148079, 0.015238627791404724, 0.0059962086379528046, -0.05849878862500191, 0.0688006803393364, -0.08256220817565918, -0.004916280508041382, -0.07917379587888718, -0.06413570046424866, 0.023247532546520233, -0.07615412771701813, -0.004007790703326464, -0.0796571671962738, -0.14724650979042053, -0.03690521419048309, 0.044062837958335876, -0.048009950667619705, -0.04650972783565521, -0.047592248767614365, -0.04840369150042534, 0.028249919414520264, -0.013071741908788681, 0.12954221665859222, -0.05163298547267914, 0.06874155253171921, -0.016369488090276718, 0.028351321816444397, 0.00481362035498023, 0.04739212617278099, -0.06346934288740158, 0.01731318049132824, -0.08225678652524948, 0.09695987403392792, -0.0917549803853035, -0.0011627463391050696, -0.11053965240716934, -0.09775421768426895, -0.008755635470151901, -0.006509967613965273, 0.08077624440193176, 0.11754442006349564, -0.13210749626159668, -0.0425269715487957, 0.12679670751094818, -0.05999647453427315, -0.061065688729286194, 0.0922454446554184, -0.031448885798454285, -0.01605350896716118, 0.04442143812775612, 0.14013226330280304, 0.08997739851474762, -0.1093715950846672, -0.02610480971634388, 0.0490739569067955, 0.04627859592437744, 0.0018941984744742513, 0.06350114941596985, -0.0008045278955250978, -0.01226184330880642, 0.02308003045618534, -0.023237785324454308, 0.03264997899532318, -0.08465246111154556, -0.08215869218111038, -0.057741399854421616, -0.08108186721801758, 0.04530856013298035, 0.01268254965543747, 0.03381390497088432, -0.06857051700353622, -0.11828240007162094, 0.10116422921419144, 0.1466180980205536, -0.07648461312055588, 0.018656494095921516, -0.06287732720375061, 0.016955381259322166, -0.02473227120935917, -0.014781026169657707, -0.19820162653923035, -0.10955855250358582, 0.04258139431476593, -0.0852029025554657, 0.042720403522253036, -0.0008424172992818058, 0.04866918921470642, 0.06317649781703949, -0.022947264835238457, -0.035362109541893005, -0.09340529143810272, -0.005673679988831282, -0.10457189381122589, -0.18612246215343475, -0.045504797250032425, -0.02541978470981121, 0.155601367354393, -0.20306091010570526, 0.010564887896180153, -0.022150229662656784, 0.1492445319890976, 0.010274413973093033, -0.04532331973314285, 0.0046164728701114655, 0.05692097544670105, -0.022598747164011, -0.0862501934170723, 0.03897128254175186, 0.016302093863487244, -0.1273123323917389, -0.045049767941236496, -0.11927222460508347, 0.034588396549224854, 0.09571244567632675, 0.022516701370477676, -0.07158894836902618, -0.0333128347992897, -0.07030696421861649, -0.05014907568693161, -0.0502292737364769, 0.021207422018051147, 0.20707574486732483, 0.008613661862909794, 0.1271543651819229, -0.06277839839458466, -0.06192139908671379, 0.007757371757179499, 0.009186974726617336, -0.0361420139670372, 0.09069014340639114, 0.09886988252401352, -0.08592447638511658, 0.08502668887376785, 0.0871363952755928, -0.07088007032871246, 0.14660894870758057, -0.04105566814541817, -0.09056057780981064, -0.006942124105989933, 0.027678538113832474, -0.026905082166194916, 0.07145407050848007, -0.11999328434467316, 0.010422374121844769, 0.023135218769311905, 0.022313276305794716, 0.07103027403354645, -0.1691489964723587, 0.01484067365527153, 0.013678832910954952, -0.038529470562934875, 0.003948397468775511, -0.015249866060912609, 0.024480972439050674, 0.08334827423095703, 0.03797183930873871, -0.007821415551006794, 0.033525243401527405, -0.0022024703212082386, -0.0776175931096077, 0.18560802936553955, -0.13910597562789917, -0.1903323084115982, -0.1212407797574997, 0.06061713024973869, -0.08872400224208832, -0.033907823264598846, 0.017079247161746025, -0.09725388884544373, -0.05471573397517204, -0.07402650266885757, -0.020651666447520256, -0.09182557463645935, 0.008987083099782467, 0.05419158190488815, 0.022773604840040207, 0.06105376407504082, -0.13697606325149536, 0.015301072970032692, -0.025871776044368744, -0.07925217598676682, -0.0076332963071763515, 0.056520331650972366, 0.08828762173652649, 0.12442037463188171, -0.01528310589492321, 0.011781741864979267, -0.025718245655298233, 0.20491868257522583, -0.06667999923229218, -0.03116905316710472, 0.10921590775251389, 0.0063876803033053875, 0.05981645733118057, 0.07686908543109894, 0.03811154142022133, -0.07680503278970718, 0.02835259959101677, 0.054862864315509796, 0.00013339027645997703, -0.26186099648475647, -0.06094061955809593, -0.04123252257704735, -0.07949963957071304, 0.09373883157968521, 0.06673607230186462, -0.008150119334459305, 0.060541048645973206, -0.022971807047724724, 0.054684389382600784, -0.04436301812529564, 0.08077781647443771, 0.094171904027462, 0.04125611111521721, 0.08648797124624252, -0.0320611372590065, -0.03055677004158497, 0.07555022835731506, 0.023628776893019676, 0.23490919172763824, -0.028242943808436394, 0.10962370038032532, 0.030171943828463554, 0.14817608892917633, -0.008414888754487038, 0.02579687349498272, 0.0182039812207222, 0.009508593939244747, 0.004078352823853493, -0.059532299637794495, -0.04158320277929306, 0.025381775572896004, 0.010747857391834259, 0.032400812953710556, -0.10473363846540451, 0.03514769300818443, 0.02002108283340931, 0.24880851805210114, -0.0037206728011369705, -0.2807617485523224, -0.06229160726070404, 0.00896789412945509, -0.029809314757585526, -0.07126452028751373, 0.018483934924006462, 0.10745570808649063, -0.12144744396209717, 0.052768152207136154, -0.04863712191581726, 0.08923503011465073, -0.0633649006485939, -0.006330363918095827, 0.035722848027944565, 0.17358346283435822, 0.007746708579361439, 0.09418625384569168, -0.21434150636196136, 0.18084384500980377, 0.018227754160761833, 0.10702883452177048, -0.07164020091295242, 0.04782425984740257, 0.01709478162229061, 0.042676281183958054, 0.0732804536819458, -0.004365983419120312, -0.06477121263742447, -0.14229555428028107, -0.09026867151260376, 0.04845759645104408, 0.1439756602048874, -0.007441554218530655, 0.07436733692884445, -0.04457324370741844, 0.014420137740671635, 0.045870233327150345, -0.09510468691587448, -0.19395291805267334, -0.1584663689136505, 0.03461238369345665, 0.05220670625567436, -0.02201095037162304, -0.06954037398099899, -0.11088074743747711, -0.04319767653942108, 0.2117283046245575, -0.03012143447995186, -0.052801188081502914, -0.14206086099147797, 0.06296749413013458, 0.14264866709709167, -0.06365698575973511, 0.022791488096117973, 0.0014565527671948075, 0.1381835788488388, 0.027227211743593216, -0.10605920106172562, 0.05790349468588829, -0.07428847998380661, -0.1484679877758026, -0.035993579775094986, 0.11851470172405243, 0.05525751784443855, 0.0330231674015522, -0.01573196053504944, 0.004233053419739008, -0.016024790704250336, -0.10023085027933121, 0.013980853371322155, 0.10186845064163208, 0.06287049502134323, 0.10170729458332062, -0.09938051551580429, 0.0010359615553170443, -0.01782951131463051, -0.02301466464996338, 0.15499745309352875, 0.19286130368709564, -0.0745394229888916, 0.08116752654314041, 0.11295318603515625, -0.08735883235931396, -0.17100082337856293, 0.052829209715127945, 0.11812371760606766, 0.016995878890156746, 0.02124819904565811, -0.2396111637353897, 0.1128142923116684, 0.11979833245277405, -0.00784103013575077, 0.06008324772119522, -0.3287430703639984, -0.11060355603694916, 0.07111047208309174, 0.11741208285093307, 0.04272463917732239, -0.12994061410427094, -0.03804568573832512, -0.04450894519686699, -0.12905389070510864, 0.1256508082151413, -0.08296245336532593, 0.12437312304973602, -0.01509249396622181, 0.07895202189683914, 0.026433801278471947, -0.037702351808547974, 0.13951949775218964, 0.004486288409680128, 0.058479972183704376, -0.04410744458436966, 0.034139569848775864, 0.020944254472851753, -0.05805432051420212, 0.053201641887426376, -0.09008666127920151, 0.07131605595350266, -0.11753862351179123, -0.02349153719842434, -0.04861769825220108, 0.08109723031520844, -0.036420587450265884, -0.05634923651814461, -0.047502804547548294, 0.031925227493047714, 0.065895676612854, -0.01817006804049015, 0.0889676883816719, 0.031063562259078026, 0.0863954946398735, 0.06421362608671188, 0.10675350576639175, -0.040620576590299606, -0.09193090349435806, -0.01282269787043333, -0.026912998408079147, 0.07493390142917633, -0.09969394654035568, 0.0166640542447567, 0.12443683296442032, 0.027733299881219864, 0.12321600317955017, 0.0561877079308033, -0.05504774674773216, 0.004164764657616615, 0.0455448180437088, -0.1178421825170517, -0.13897791504859924, -0.03543245792388916, -0.024507224559783936, -0.11962258815765381, 0.03556725010275841, 0.10866694152355194, -0.04785499721765518, -0.01647106185555458, -0.020357593894004822, 0.02090725302696228, -0.012375960126519203, 0.15150180459022522, 0.03175900876522064, 0.046859871596097946, -0.08534860610961914, 0.14321137964725494, 0.07470767945051193, -0.10103772580623627, 0.08116991817951202, 0.10153828561306, -0.07595127820968628, -0.02072964236140251, 0.06986239552497864, 0.147535502910614, -0.04883250221610069, -0.060297224670648575, -0.07952187955379486, -0.09414482116699219, 0.05189204961061478, 0.06888838857412338, 0.03806568682193756, -0.0008801845251582563, -0.03636377677321434, 0.04154450073838234, -0.15346002578735352, 0.07263961434364319, 0.03901681303977966, 0.07078395783901215, -0.15492987632751465, 0.10062158852815628, 0.0253162682056427, 0.042937714606523514, -0.016291063278913498, 0.0011253205593675375, -0.102586530148983, -0.02157430164515972, -0.1377524733543396, -0.0049951206892728806, -0.03937321901321411, 0.014963412657380104, -0.008328309282660484, -0.03219029679894447, -0.05123168230056763, 0.05566401407122612, -0.06585534662008286, -0.06696902960538864, -0.0002872512850444764, 0.06318429112434387, -0.12253621220588684, 0.012963352724909782, 0.006634782999753952, -0.09975486248731613, 0.07949846237897873, 0.061883069574832916, 0.016914747655391693, 0.031234467402100563, -0.1470668762922287, -0.019237875938415527, 0.032800931483507156, 0.0341971330344677, 0.04804132878780365, -0.0727081149816513, 0.001604636781848967, -0.005394958425313234, 0.04868937283754349, -0.003805170301347971, 0.05561643838882446, -0.12828437983989716, -0.041293878108263016, -0.07961840182542801, -0.04278762638568878, -0.08343847841024399, 0.05265524238348007, 0.1137513592839241, 0.03959314897656441, 0.15630824863910675, -0.07335226237773895, 0.03008372336626053, -0.16554443538188934, -0.016750432550907135, -0.0075095598585903645, -0.029193812981247902, -0.05989283695816994, -0.03272201865911484, 0.08127864450216293, -0.05393782630562782, 0.14175903797149658, -0.021062318235635757, 0.06357862800359726, 0.023085182532668114, -0.039506491273641586, -0.02418370172381401, 0.013387667946517467, 0.21486809849739075, 0.08493192493915558, -0.009039702825248241, 0.07511135935783386, -0.0060802483931183815, 0.07428120076656342, 0.042263127863407135, 0.172538623213768, 0.13531450927257538, -0.017084600403904915, 0.09474925696849823, 0.07103551924228668, -0.10913695394992828, -0.1327979415655136, 0.1185174509882927, -0.03655830770730972, 0.12034787237644196, -0.04483669996261597, 0.17277103662490845, 0.08005291223526001, -0.14280864596366882, 0.0318845696747303, -0.029528887942433357, -0.09690864384174347, -0.1175217553973198, -0.026938648894429207, -0.08457677811384201, -0.13305573165416718, 0.01728852279484272, -0.12154651433229446, 0.02214362844824791, 0.04463978111743927, 0.016121475026011467, 0.009996306151151657, 0.11203289777040482, 0.0033048747573047876, -0.006933530792593956, 0.06008201092481613, 0.009062934666872025, -0.002483412390574813, -0.05864261835813522, -0.09373849630355835, 0.04447720944881439, 0.012480652891099453, 0.08830810338258743, -0.03741157054901123, -0.015703843906521797, 0.05896243080496788, -0.021246349439024925, -0.06125127896666527, 0.025915291160345078, 0.005848430097103119, 0.02933979034423828, 0.054827023297548294, 0.04020829126238823, -0.03344876691699028, -0.050835706293582916, 0.25111210346221924, -0.06971227377653122, -0.06515911221504211, -0.11491299420595169, 0.17692439258098602, 0.027858443558216095, -0.013025177642703056, 0.06754563748836517, -0.10397719591856003, -0.046539656817913055, 0.1635621190071106, 0.1445709764957428, -0.0606500618159771, -0.025890246033668518, 0.007381260395050049, -0.01812603510916233, -0.056442953646183014, 0.136391282081604, 0.11985919624567032, 0.07093552500009537, -0.0494137667119503, -0.0105929896235466, -0.0164518803358078, -0.02062160149216652, -0.10065034031867981, 0.04950819909572601, 0.012220011092722416, -0.011094546876847744, -0.0320306196808815, 0.05914059653878212, -0.02909787930548191, -0.1382295936346054, 0.01295194961130619, -0.13290920853614807, -0.1684482842683792, -0.03710947185754776, 0.10266689956188202, -0.02184159867465496, 0.04769781231880188, -0.0029205093160271645, -0.015675971284508705, 0.10971052199602127, -0.017960168421268463, -0.06839945912361145, -0.07587962597608566, 0.0638507753610611, -0.0452476367354393, 0.20398542284965515, 0.004646057263016701, 0.08100634813308716, 0.10770578682422638, 0.04554111137986183, -0.1326245814561844, 0.03315424546599388, 0.06119205802679062, -0.07647865265607834, 0.05838733911514282, 0.13739797472953796, -0.05399416387081146, 0.06754285842180252, 0.04074883088469505, -0.08238302171230316, -0.018988246098160744, -0.07272457331418991, -0.018543563783168793, -0.06500761210918427, -0.008523307740688324, -0.06848219037055969, 0.15677274763584137, 0.18938663601875305, -0.030595483258366585, -0.017349116504192352, -0.07558764517307281, 0.03169560804963112, 0.033654503524303436, 0.11118608713150024, -0.018498899415135384, -0.21732191741466522, 0.00888350885361433, 0.040678080171346664, 0.04266383498907089, -0.22287029027938843, -0.07539297640323639, 0.023480311036109924, -0.042328234761953354, -0.0748230367898941, 0.09356813132762909, 0.07641927897930145, 0.027167560532689095, -0.0411091148853302, -0.13106484711170197, -0.050048183649778366, 0.13016031682491302, -0.1481027901172638, -0.04626721888780594 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlnet_roberta_summarization_cnn_dailymail This model is a fine-tuned version of [](https://huggingface.co/) on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["cnn_dailymail"], "model-index": [{"name": "xlnet_roberta_summarization_cnn_dailymail", "results": []}]}
text2text-generation
Ayham/xlnet_roberta_summarization_cnn_dailymail
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "dataset:cnn_dailymail", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us
# xlnet_roberta_summarization_cnn_dailymail This model is a fine-tuned version of [](URL on the cnn_dailymail dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
[ "# xlnet_roberta_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n", "# xlnet_roberta_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ 62, 41, 6, 12, 8, 3, 118, 4, 38 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #dataset-cnn_dailymail #autotrain_compatible #endpoints_compatible #region-us \n# xlnet_roberta_summarization_cnn_dailymail\n\nThis model is a fine-tuned version of [](URL on the cnn_dailymail dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.18.3\n- Tokenizers 0.10.3" ]
[ -0.08226434141397476, 0.13918323814868927, -0.0038389605470001698, 0.07111088931560516, 0.13159547746181488, 0.04041111096739769, 0.11423606425523758, 0.12152039259672165, -0.12134427577257156, 0.05610841512680054, 0.07135500758886337, 0.0651995986700058, 0.048238370567560196, 0.15607024729251862, -0.027741564437747, -0.22364388406276703, 0.003118420485407114, -0.006062682252377272, -0.07723603397607803, 0.1182486042380333, 0.08493904769420624, -0.09923292696475983, 0.06002544239163399, -0.0029377180617302656, -0.13348987698554993, 0.02090582251548767, -0.04282020032405853, -0.05283229053020477, 0.09660092741250992, -0.013708176091313362, 0.09327033907175064, 0.02873941883444786, 0.14238223433494568, -0.2056555300951004, 0.0040532308630645275, 0.09681028872728348, 0.03484060615301132, 0.09221678972244263, 0.06382045894861221, -0.012260650284588337, 0.13878794014453888, -0.16703522205352783, 0.10101781785488129, 0.010347220115363598, -0.07811211794614792, -0.10687002539634705, -0.10075840353965759, 0.05850179120898247, 0.0926986038684845, 0.11727555841207504, 0.006359136663377285, 0.1375931203365326, -0.10682450979948044, 0.08366350084543228, 0.16805408895015717, -0.20821337401866913, -0.053810544312000275, 0.06945521384477615, 0.02386586181819439, 0.060481421649456024, -0.0901012048125267, -0.015599061734974384, 0.022400764748454094, 0.023721087723970413, 0.09887422621250153, 0.00020324913202784956, -0.10257649421691895, 0.0017186043551191688, -0.12727683782577515, -0.025133969262242317, 0.08162079006433487, 0.01685231551527977, -0.007530578412115574, -0.09627338498830795, -0.05544013902544975, -0.14146575331687927, -0.02208421565592289, -0.01788703165948391, 0.040249139070510864, -0.05650307238101959, -0.06751354783773422, -0.0454254075884819, -0.07008418440818787, -0.04837791249155998, -0.027335837483406067, 0.11461988836526871, 0.045118484646081924, 0.012777489610016346, -0.05296452343463898, 0.11808464676141739, 0.07030945271253586, -0.12449070811271667, 0.01125689223408699, 0.014261600561439991, -0.09407082945108414, -0.05552341789007187, -0.024444177746772766, -0.033897947520017624, -0.009627891704440117, 0.13473345339298248, -0.07506763935089111, 0.0781012549996376, 0.02337823621928692, -0.01677859015762806, -0.010897117666900158, 0.13563121855258942, -0.04823868349194527, -0.042198844254016876, -0.012581194750964642, 0.1069454625248909, 0.007494064513593912, -0.01020012330263853, -0.06656711548566818, 0.0014849593862891197, 0.07168207317590714, 0.06285787373781204, -0.034523360431194305, 0.04377591237425804, -0.01476603839546442, -0.019901959225535393, 0.04051226004958153, -0.13683122396469116, 0.04236087575554848, 0.0145006338134408, -0.10333552956581116, -0.007804983761161566, 0.013425126671791077, 0.0001845283550210297, -0.04960044100880623, 0.13098929822444916, -0.07833658158779144, 0.0113992840051651, -0.0801805853843689, -0.07071822136640549, 0.01735231652855873, -0.09823019802570343, -0.04113820567727089, -0.03296240046620369, -0.1931362748146057, -0.05948089063167572, 0.027932219207286835, -0.05758535861968994, -0.0357929952442646, -0.057996731251478195, -0.07511751353740692, 0.01918550208210945, -0.006491976324468851, 0.18443970382213593, -0.06798695772886276, 0.06445609778165817, -0.02656552754342556, 0.024420814588665962, 0.06947696208953857, 0.04159769043326378, -0.07026752084493637, -0.007426326163113117, -0.09601055830717087, 0.08792915940284729, -0.0736839696764946, 0.004818808753043413, -0.10757730156183243, -0.0954722911119461, 0.016993893310427666, -0.013272655196487904, 0.05211425572633743, 0.1354384869337082, -0.20284481346607208, -0.03163699060678482, 0.1194465309381485, -0.05840887874364853, -0.009797747246921062, 0.05326395109295845, -0.060100819915533066, -0.000885547196958214, 0.06735187768936157, 0.12607333064079285, 0.06487526744604111, -0.12767136096954346, -0.009958975948393345, -0.008330589160323143, 0.02903510257601738, 0.006709394510835409, 0.018393810838460922, 0.011855656281113625, 0.0672236829996109, 0.0037510038819164038, -0.039778560400009155, 0.00601933803409338, -0.09459864348173141, -0.07040747255086899, -0.03758048266172409, -0.07131752371788025, 0.012280842289328575, 0.02660827524960041, 0.03738957270979881, -0.05591276288032532, -0.10972423106431961, 0.09108529984951019, 0.12666992843151093, -0.05256981402635574, 0.015338878147304058, -0.07117818295955658, -0.041961174458265305, 0.009188096038997173, -0.023615093901753426, -0.19710928201675415, -0.11294892430305481, 0.021300900727510452, -0.06358436495065689, 0.043441612273454666, -0.010726355947554111, 0.0666419267654419, 0.04020495340228081, -0.03820919990539551, -0.014941093511879444, -0.06168549135327339, -0.000417279516113922, -0.0895567387342453, -0.20659756660461426, -0.029690023511648178, -0.018401023000478745, 0.1561717987060547, -0.2154768407344818, 0.015293345786631107, -0.019013533368706703, 0.14944015443325043, 0.02435445785522461, -0.05744148790836334, 0.02056802250444889, 0.05161784961819649, 0.007472275756299496, -0.09397619217634201, 0.028589673340320587, -0.01025962084531784, -0.07211262732744217, -0.02013069950044155, -0.13770706951618195, -0.029288481920957565, 0.07765599340200424, 0.08930975943803787, -0.10409698635339737, 0.018981432542204857, -0.0739608034491539, -0.04976553097367287, -0.09187132120132446, 0.03631135821342468, 0.2047753781080246, 0.035059429705142975, 0.1263345330953598, -0.05027918517589569, -0.08474504202604294, -0.009761958383023739, 0.023809034377336502, 0.00717310793697834, 0.09867369383573532, 0.07938346266746521, -0.047210920602083206, 0.07351259142160416, 0.011781202629208565, -0.04802992194890976, 0.1362357884645462, -0.04587511718273163, -0.08682983368635178, -0.0064037092961370945, -0.026897914707660675, -0.024319468066096306, 0.0992937982082367, -0.03260355442762375, 0.005495138932019472, 0.03150563687086105, 0.026290562003850937, 0.03774029761552811, -0.18373842537403107, -0.0031706644222140312, 0.011582357808947563, -0.060262877494096756, -0.03081781230866909, -0.016330596059560776, 0.053452327847480774, 0.10226046293973923, 0.0030863219872117043, -0.033639851957559586, 0.011457053013145924, -0.008191931992769241, -0.07260287553071976, 0.1760420799255371, -0.11419498920440674, -0.16807788610458374, -0.08233991265296936, 0.03182540461421013, -0.0342949740588665, -0.054117053747177124, 0.0005109147168695927, -0.10980748385190964, -0.06816624104976654, -0.11381314694881439, -0.023491540923714638, 0.005726280622184277, -0.0013465327210724354, 0.0359363928437233, 0.017829526215791702, 0.0474005863070488, -0.1369961053133011, 0.015834389254450798, -0.05216333642601967, -0.07906311005353928, 0.01955420896410942, 0.07714914530515671, 0.05888395011425018, 0.1479993313550949, -0.01494663581252098, 0.029662594199180603, -0.016025716438889503, 0.19414274394512177, -0.09359924495220184, 0.01630830205976963, 0.10169543325901031, 0.018846049904823303, 0.04414478689432144, 0.12084508687257767, 0.03951307758688927, -0.07185474038124084, 0.025829195976257324, 0.09204324334859848, -0.020698867738246918, -0.2653106153011322, -0.0633934959769249, -0.017042746767401695, -0.08106498420238495, 0.0856945812702179, 0.059644460678100586, 0.0006011977093294263, 0.008228333666920662, -0.008847330696880817, -0.01650463603436947, -0.00789861660450697, 0.05457717552781105, 0.08073031902313232, 0.058599747717380524, 0.08848249167203903, -0.032810457050800323, -0.031912874430418015, 0.06434520334005356, 0.0007453401340171695, 0.2222391813993454, -0.059150464832782745, 0.07466872036457062, 0.018593139946460724, 0.09155239164829254, -0.017039118334650993, 0.030849434435367584, 0.02601371705532074, -0.01000108290463686, 0.01627667061984539, -0.051986370235681534, 0.0025522925425320864, 0.003723784815520048, -0.015170219354331493, -0.0031023661140352488, -0.05355549231171608, 0.03759700059890747, 0.018619578331708908, 0.27699533104896545, 0.03526323288679123, -0.2849409580230713, -0.04528786987066269, -0.02201538160443306, -0.03402072191238403, -0.045986514538526535, -0.014234984293580055, 0.09433180093765259, -0.11551891267299652, 0.07051357626914978, -0.04867691919207573, 0.08326084166765213, -0.07644357532262802, -0.005152484402060509, 0.06370154023170471, 0.13822905719280243, -0.018668610602617264, 0.060477420687675476, -0.21555495262145996, 0.21938881278038025, 0.011308805085718632, 0.12881703674793243, -0.06948835402727127, 0.02135251834988594, 0.016670558601617813, 0.041374705731868744, 0.06611894816160202, -0.005906450096517801, -0.030065447092056274, -0.13977672159671783, -0.13024909794330597, 0.02482193149626255, 0.1095307469367981, 0.0016553077148273587, 0.07909587770700455, -0.015793576836586, -0.016071483492851257, 0.03730681911110878, -0.1589948534965515, -0.15786902606487274, -0.11607950180768967, 0.036926548928022385, 0.03513214737176895, -0.04644935950636864, -0.05214167758822441, -0.12067658454179764, -0.02703786827623844, 0.17518135905265808, 0.025518083944916725, -0.062445104122161865, -0.14182564616203308, 0.05723848566412926, 0.15509818494319916, -0.039770349860191345, 0.009953156113624573, 0.036737408488988876, 0.11447315663099289, 0.043318863958120346, -0.08501765877008438, 0.030029717832803726, -0.059262458235025406, -0.18307051062583923, -0.05529532954096794, 0.13020919263362885, 0.08895845711231232, 0.04682512953877449, -0.020664645358920097, 0.03138814494013786, -0.0036195688880980015, -0.07914812117815018, 0.0017159964190796018, 0.07567603886127472, 0.08425571024417877, 0.06220075860619545, -0.06324508786201477, -0.009640287607908249, -0.04745710268616676, -0.02306361310184002, 0.09859562665224075, 0.18333646655082703, -0.0764109343290329, 0.10283951461315155, 0.06028258055448532, -0.06809906661510468, -0.1673760563135147, 0.056610047817230225, 0.12737102806568146, 0.010218929499387741, 0.03630758449435234, -0.20148038864135742, 0.11733057349920273, 0.12330871820449829, -0.011763008311390877, 0.05330131575465202, -0.35613542795181274, -0.12729595601558685, 0.04932485148310661, 0.08226045221090317, 0.012480131350457668, -0.1070680245757103, -0.020551415160298347, -0.0052060955204069614, -0.12751425802707672, 0.14881631731987, -0.06888499110937119, 0.10899748653173447, -0.011073365807533264, 0.11325710266828537, 0.024544626474380493, -0.047627970576286316, 0.11070071905851364, 0.07251113653182983, 0.07296067476272583, -0.04734383895993233, 0.010535847395658493, 0.045233212411403656, -0.06902781873941422, 0.046123307198286057, -0.044342849403619766, 0.06833315640687943, -0.12933005392551422, -0.012376520782709122, -0.08302255719900131, 0.043535780161619186, -0.04609490558505058, -0.04341728985309601, -0.03998703882098198, 0.05575684830546379, 0.07296331226825714, -0.04479086399078369, 0.07415149360895157, 0.01121761929243803, 0.081793874502182, 0.07556957751512527, 0.08491694182157516, -0.03947470337152481, -0.09677603095769882, -0.00774754723533988, -0.007181351538747549, 0.0314386785030365, -0.12056685984134674, 0.0194193497300148, 0.1360286921262741, 0.04897232726216316, 0.122447669506073, 0.030769797042012215, -0.04138612002134323, -0.022795040160417557, 0.03698519244790077, -0.12717051804065704, -0.08585204184055328, 0.031882401555776596, -0.07036467641592026, -0.10816178470849991, 0.014152697287499905, 0.09997532516717911, -0.037854086607694626, -0.01820426806807518, 0.006119864527136087, 0.031139113008975983, 0.010150295682251453, 0.21926723420619965, 0.026838799938559532, 0.06157932057976723, -0.11340110749006271, 0.12672826647758484, 0.06835468858480453, -0.08422885835170746, 0.04474440962076187, 0.13076895475387573, -0.09759677201509476, -0.012902937829494476, 0.10408784449100494, 0.14558471739292145, -0.03430137783288956, -0.033699970692396164, -0.09216674417257309, -0.08752213418483734, 0.07764604687690735, 0.12649206817150116, 0.032205525785684586, 0.007308754604309797, -0.053832974284887314, 0.010711169801652431, -0.14629609882831573, 0.06681478768587112, 0.08793269097805023, 0.05671275779604912, -0.08274554461240768, 0.14299795031547546, 0.025541329756379128, 0.03566794469952583, -0.017287131398916245, 0.012739882804453373, -0.06686974316835403, -0.009472941048443317, -0.07652207463979721, -0.008149534463882446, -0.014921538531780243, -0.004224738106131554, -0.01981303095817566, -0.050052106380462646, -0.04405120015144348, 0.04214400798082352, -0.07189847528934479, -0.05425082519650459, -0.00527831818908453, 0.03781881555914879, -0.13643716275691986, -0.0003325871075503528, -0.000541465065907687, -0.09129755944013596, 0.07015533000230789, 0.04890820011496544, -0.0038032527081668377, 0.029254291206598282, -0.14724549651145935, -0.032244812697172165, 0.030698025599122047, 0.017364466562867165, 0.07347914576530457, -0.07636461406946182, -0.0024965405464172363, -0.017322693020105362, 0.037187982350587845, 0.01728973351418972, 0.04355742409825325, -0.1140015572309494, 0.013494393788278103, -0.06927483528852463, -0.03209984302520752, -0.06700021773576736, 0.03888848051428795, 0.11030688136816025, 0.038659315556287766, 0.17781217396259308, -0.07736171036958694, 0.03902618959546089, -0.18293875455856323, -0.029975226148962975, -0.0033393604680895805, -0.04092231020331383, -0.06434111297130585, -0.011847390793263912, 0.1089726984500885, -0.05363207310438156, 0.12641409039497375, 0.015811972320079803, 0.06127358600497246, 0.0396222248673439, -0.024224156513810158, -0.06834334880113602, 0.01270620059221983, 0.1404561549425125, 0.06165105104446411, -0.019416000694036484, 0.11560338735580444, -0.013891945593059063, 0.04624840244650841, 0.04503600299358368, 0.2272011786699295, 0.12513630092144012, 0.02814488857984543, 0.07179941236972809, 0.054572898894548416, -0.13389703631401062, -0.1112663671374321, 0.105157770216465, -0.09046798199415207, 0.11896193772554398, -0.06336380541324615, 0.19563335180282593, 0.059646278619766235, -0.14813193678855896, 0.05424380674958229, -0.02161048911511898, -0.10294154286384583, -0.1023060530424118, -0.01677720807492733, -0.07616107910871506, -0.10054291039705276, 0.033425576984882355, -0.11166266351938248, 0.06278296560049057, 0.09464836120605469, 0.03417845070362091, 0.020663172006607056, 0.11158493161201477, -0.00043807175825349987, -0.002364225685596466, 0.06059647724032402, 0.025042347609996796, 0.0009675332112237811, -0.06625274568796158, -0.06247856840491295, 0.021015722304582596, 0.027292827144265175, 0.08526802808046341, -0.03631764277815819, 0.00494052330031991, 0.0451083704829216, -0.01958589442074299, -0.0736030861735344, 0.03876035660505295, 0.0024850580375641584, 0.06748015433549881, 0.05773705989122391, 0.05914099141955376, 0.0072272764518857, -0.046259935945272446, 0.27717819809913635, -0.057666778564453125, -0.1123083084821701, -0.12348056584596634, 0.21964405477046967, 0.04909698665142059, -0.02742144465446472, 0.0623607262969017, -0.10630614310503006, -0.04483361542224884, 0.1564481556415558, 0.17452332377433777, -0.053590621799230576, -0.026067214086651802, -0.031523577868938446, -0.011549847200512886, -0.021161599084734917, 0.1161315068602562, 0.10133881121873856, 0.057203371077775955, -0.04912680387496948, -0.020186755806207657, -0.0190399382263422, -0.03785209357738495, -0.07592884451150894, 0.06086963415145874, 0.019375324249267578, -0.007034608628600836, -0.03217059001326561, 0.07786213606595993, -0.01984843797981739, -0.20101101696491241, 0.040887705981731415, -0.17053021490573883, -0.18828372657299042, -0.003419888438656926, 0.10147920995950699, -0.025950005277991295, 0.045903053134679794, 0.014506842941045761, -0.016036072745919228, 0.0924653485417366, -0.030966026708483696, -0.03734344244003296, -0.10763725638389587, 0.06839999556541443, -0.11921484023332596, 0.224534809589386, -0.005079156253486872, 0.07766090333461761, 0.08950896561145782, 0.019060872495174408, -0.12704528868198395, 0.06042678281664848, 0.04486408829689026, -0.0909445658326149, 0.03165512531995773, 0.15926562249660492, -0.04165089502930641, 0.03282792493700981, 0.036906421184539795, -0.10477238893508911, -0.0072640310972929, -0.07845186442136765, -0.020309535786509514, -0.06781186908483505, -0.014649169519543648, -0.025794534012675285, 0.1510917991399765, 0.21045444905757904, -0.00814974308013916, 0.03264572471380234, -0.07926198840141296, -0.005530450493097305, 0.02325405552983284, 0.08003869652748108, -0.028930850327014923, -0.22442622482776642, 0.03370332717895508, 0.007627095561474562, 0.039847563952207565, -0.1735554337501526, -0.07243362069129944, 0.010406631976366043, -0.07403359562158585, -0.051881518214941025, 0.10674164444208145, 0.0443531796336174, 0.03981500491499901, -0.021028753370046616, -0.07394122332334518, -0.009934711270034313, 0.14453929662704468, -0.19132137298583984, -0.038719434291124344 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlnetgpt2_xsum7 This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "xlnetgpt2_xsum7", "results": []}]}
text2text-generation
Ayham/xlnetgpt2_xsum7
[ "transformers", "pytorch", "tensorboard", "encoder-decoder", "text2text-generation", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
# xlnetgpt2_xsum7 This model is a fine-tuned version of [](URL on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 3.0 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# xlnetgpt2_xsum7\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n", "# xlnetgpt2_xsum7\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ 53, 33, 6, 12, 8, 3, 118, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #encoder-decoder #text2text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n# xlnetgpt2_xsum7\n\nThis model is a fine-tuned version of [](URL on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 2000\n- num_epochs: 3.0\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ -0.08108361065387726, 0.14206042885780334, -0.003036821261048317, 0.06268417090177536, 0.1430549919605255, 0.04181598499417305, 0.11200176179409027, 0.12846826016902924, -0.09789770096540451, 0.08584801107645035, 0.08498413115739822, 0.06462015211582184, 0.06253357231616974, 0.1187276691198349, -0.02180410362780094, -0.2701355516910553, 0.02378641627728939, -0.0031274990178644657, -0.08146043866872787, 0.08807922154664993, 0.08605767786502838, -0.097637839615345, 0.07185417413711548, 0.02135634981095791, -0.15965573489665985, 0.022927111014723778, -0.05363864451646805, -0.07670822739601135, 0.07934806495904922, 0.01835094764828682, 0.08947324007749557, 0.009129445999860764, 0.09817469865083694, -0.22984233498573303, 0.0039586559869349, 0.06950218975543976, 0.02587246708571911, 0.08824360370635986, 0.08608022332191467, -0.0038609292823821306, 0.16235294938087463, -0.14877554774284363, 0.09956212341785431, 0.02853880263864994, -0.053663481026887894, -0.13199515640735626, -0.07500191777944565, 0.07656579464673996, 0.0646209716796875, 0.10106553882360458, 0.013833423145115376, 0.14117014408111572, -0.09411285072565079, 0.0722304955124855, 0.18903608620166779, -0.2470838725566864, -0.06511840969324112, 0.009688880294561386, 0.06578774005174637, 0.06424111872911453, -0.10630816221237183, -0.014714282937347889, 0.036751456558704376, 0.03614068403840065, 0.07272236049175262, 0.008056235499680042, -0.03249392658472061, -0.0024935801047831774, -0.13395540416240692, -0.020099777728319168, 0.1486377865076065, 0.051497314125299454, -0.03318241983652115, -0.11454849690198898, -0.046882253140211105, -0.09954019635915756, -0.017875779420137405, -0.03424540534615517, 0.02529834769666195, -0.03475647047162056, -0.0746828019618988, -0.05453355237841606, -0.06822629272937775, -0.05029444769024849, -0.003696885658428073, 0.16551169753074646, 0.054976508021354675, 0.00974609237164259, -0.04288048297166824, 0.08622332662343979, 0.08230510354042053, -0.12106284499168396, -0.010535645298659801, -0.013428669422864914, -0.0653170496225357, -0.04769987240433693, -0.052416976541280746, -0.045714639127254486, 0.028337860479950905, 0.12474144995212555, -0.06329534202814102, 0.09105546772480011, 0.02045566402375698, 0.005926282610744238, 0.000547398638445884, 0.12283802032470703, -0.042026955634355545, -0.024427171796560287, -0.010065628215670586, 0.09080769866704941, 0.018587976694107056, -0.020107749849557877, -0.08366710692644119, -0.022188914939761162, 0.08884003758430481, 0.060082562267780304, -0.024515505880117416, 0.021760469302535057, -0.042984358966350555, -0.0380796417593956, 0.006253204774111509, -0.12274383753538132, 0.05368538200855255, 0.010561857372522354, -0.08729700744152069, 0.043553426861763, 0.017066339030861855, 0.0005720392218790948, -0.04730253294110298, 0.0874490737915039, -0.08722899854183197, -0.004553414881229401, -0.07956835627555847, -0.0602155365049839, 0.03573028743267059, -0.05847909301519394, -0.00625659991055727, -0.07354200631380081, -0.147990420460701, -0.033824216574430466, 0.049124594777822495, -0.05153060704469681, -0.0402396135032177, -0.05139205604791641, -0.05167699605226517, 0.04458686709403992, -0.0012843289878219366, 0.11369556188583374, -0.04798221215605736, 0.06889297068119049, -0.01579131931066513, 0.032695185393095016, 0.026471540331840515, 0.0452306903898716, -0.07178792357444763, 0.01656213589012623, -0.09456458687782288, 0.08117912709712982, -0.09035605937242508, 0.014286625199019909, -0.1107674315571785, -0.09818173944950104, -0.015121033415198326, -0.011827087961137295, 0.07466469705104828, 0.12770894169807434, -0.13636763393878937, -0.05238962545990944, 0.13081656396389008, -0.0769822970032692, -0.06895677000284195, 0.09032120555639267, -0.036282625049352646, -0.015094229951500893, 0.04141136631369591, 0.12639884650707245, 0.0879666656255722, -0.11225444823503494, -0.034040823578834534, 0.020003482699394226, 0.049511928111314774, 0.0076594785787165165, 0.06574521213769913, -0.009776918217539787, 0.0011681712931022048, 0.01938130147755146, -0.009023981168866158, 0.020160771906375885, -0.08586435765028, -0.08967922627925873, -0.05560696870088577, -0.07023569196462631, 0.03759785369038582, 0.019653869792819023, 0.042451296001672745, -0.06707853078842163, -0.1211942508816719, 0.11181941628456116, 0.1383630782365799, -0.06584345549345016, 0.023761050775647163, -0.06425147503614426, 0.01744651421904564, -0.034460004419088364, -0.018804892897605896, -0.19551117718219757, -0.11329640448093414, 0.04436716437339783, -0.09914837777614594, 0.03688301518559456, -0.007787852548062801, 0.06130925938487053, 0.06962987035512924, -0.02412131242454052, -0.04138163477182388, -0.1019505113363266, -0.0034632051829248667, -0.09283365309238434, -0.20000198483467102, -0.0461864210665226, -0.03309006616473198, 0.14009560644626617, -0.2020578533411026, 0.005338259972631931, -0.011149651370942593, 0.16392123699188232, 0.02059910260140896, -0.05100555717945099, 0.006522458046674728, 0.05650221183896065, -0.021089641377329826, -0.08362463861703873, 0.03789044916629791, 0.01025826670229435, -0.12110608071088791, -0.056403059512376785, -0.11780592799186707, 0.04569508135318756, 0.09655606746673584, 0.03417467325925827, -0.07625961303710938, -0.02743934839963913, -0.07467053830623627, -0.0534539520740509, -0.06093272194266319, 0.02642885409295559, 0.19010232388973236, 0.014838357456028461, 0.11827800422906876, -0.055789075791835785, -0.061046890914440155, 0.014874055981636047, 0.009762422181665897, -0.04794946312904358, 0.08562831580638885, 0.09564626216888428, -0.09417138248682022, 0.0868912786245346, 0.0851837694644928, -0.06297969818115234, 0.15691642463207245, -0.04809011518955231, -0.10103017091751099, -0.018002057448029518, 0.024233810603618622, -0.009626076556742191, 0.0810251235961914, -0.1102418303489685, 0.00456761522218585, 0.023426253348588943, 0.026143312454223633, 0.06060032919049263, -0.1639665961265564, 0.009030263870954514, 0.018478835001587868, -0.0297255702316761, -0.0060036927461624146, -0.00551189761608839, 0.021897289901971817, 0.07881169766187668, 0.04372958466410637, -0.0030518362764269114, 0.030858715996146202, -0.004619767423719168, -0.08248896896839142, 0.1802133172750473, -0.14182861149311066, -0.17136316001415253, -0.12635503709316254, 0.055905669927597046, -0.08113276213407516, -0.03423541784286499, 0.02568848431110382, -0.09816121309995651, -0.05511575937271118, -0.07158031314611435, -0.02504560723900795, -0.09642452746629715, 0.018788160756230354, 0.06506162136793137, 0.021587833762168884, 0.06230628117918968, -0.1330098658800125, 0.014320321381092072, -0.023742053657770157, -0.06880174577236176, -0.0073228869587183, 0.058587852865457535, 0.0963498130440712, 0.11131075024604797, -0.00470517436042428, 0.019832979887723923, -0.02782241627573967, 0.2076561450958252, -0.06596660614013672, -0.02655922621488571, 0.10119537264108658, 0.007278153672814369, 0.07019713521003723, 0.0854305699467659, 0.044194627553224564, -0.07244064658880234, 0.028977248817682266, 0.05354222282767296, 0.002409064443781972, -0.25394997000694275, -0.0505199171602726, -0.03999776765704155, -0.0731697604060173, 0.09776283800601959, 0.06897720694541931, -0.005508297123014927, 0.052935224026441574, -0.019461598247289658, 0.04481704533100128, -0.04522869363427162, 0.09022505581378937, 0.08891379833221436, 0.05747544765472412, 0.08216895163059235, -0.03466274216771126, -0.032817840576171875, 0.07006660103797913, 0.001441099215298891, 0.2522803843021393, -0.007601732853800058, 0.10842975229024887, 0.03047678805887699, 0.13747744262218475, -0.006347671616822481, 0.03489065542817116, 0.024906296283006668, 0.008835348300635815, 0.008008785545825958, -0.06745939701795578, -0.03553672134876251, 0.034696754068136215, 0.011006031185388565, 0.05247727036476135, -0.11454693228006363, 0.044266268610954285, 0.022015884518623352, 0.2475610375404358, 0.001300928764976561, -0.28141066431999207, -0.07064546644687653, 0.010816468857228756, -0.03684864938259125, -0.06049656122922897, 0.010786404833197594, 0.08706853538751602, -0.1271793097257614, 0.06463603675365448, -0.06143588572740555, 0.08635144680738449, -0.05600908398628235, -0.012502040714025497, 0.040875256061553955, 0.16074539721012115, 0.0015642275102436543, 0.08657737076282501, -0.20737996697425842, 0.2019433081150055, 0.016150515526533127, 0.11899441480636597, -0.06850789487361908, 0.053136155009269714, 0.025224696844816208, 0.03710402548313141, 0.08179228752851486, -0.007386103738099337, -0.09859970957040787, -0.1464555412530899, -0.09088245034217834, 0.04002578556537628, 0.1441987156867981, -0.011246350593864918, 0.0829351469874382, -0.04041925072669983, 0.008616502396762371, 0.04624829441308975, -0.07059412449598312, -0.19327962398529053, -0.1533990204334259, 0.03800663352012634, 0.029303984716534615, -0.022607294842600822, -0.0744432657957077, -0.11927594244480133, -0.02865569107234478, 0.1861933469772339, -0.018256306648254395, -0.04272704944014549, -0.15446555614471436, 0.050928592681884766, 0.1437697857618332, -0.06970148533582687, 0.028384309262037277, 0.000502720067743212, 0.14031831920146942, 0.026940403506159782, -0.10401102155447006, 0.05686862766742706, -0.07628923654556274, -0.15980178117752075, -0.04787345603108406, 0.11565742641687393, 0.056034039705991745, 0.03744688630104065, -0.017772618681192398, 0.02003570832312107, -0.011968583799898624, -0.09463471174240112, 0.028404127806425095, 0.11466148495674133, 0.0640137791633606, 0.07661525905132294, -0.09142417460680008, 0.008241243660449982, -0.017054496333003044, -0.0210803784430027, 0.14153383672237396, 0.19984839856624603, -0.07422589510679245, 0.08897443115711212, 0.10341271758079529, -0.08644305169582367, -0.15978974103927612, 0.055408623069524765, 0.11711202561855316, 0.015369405038654804, 0.02417641133069992, -0.23422366380691528, 0.11261434853076935, 0.10935104638338089, -0.008755341172218323, 0.05097315460443497, -0.3261040449142456, -0.10824573785066605, 0.06307791918516159, 0.11667986959218979, 0.035853371024131775, -0.13659437000751495, -0.04215266928076744, -0.03470849245786667, -0.12191983312368393, 0.11449727416038513, -0.07386637479066849, 0.11593718826770782, -0.022344626486301422, 0.0821474939584732, 0.021401723846793175, -0.04215226322412491, 0.13488492369651794, 0.00517634954303503, 0.05885728448629379, -0.04491102322936058, 0.023432457819581032, 0.034574758261442184, -0.06636638939380646, 0.05259581655263901, -0.07363786548376083, 0.06344489753246307, -0.1369418501853943, -0.018760282546281815, -0.053782325237989426, 0.0824500024318695, -0.04592429846525192, -0.05435824766755104, -0.04082120582461357, 0.03692714497447014, 0.07940756529569626, -0.023452026769518852, 0.10081885755062103, 0.04143378511071205, 0.09028501063585281, 0.07136596739292145, 0.10214263945817947, -0.01580299809575081, -0.1011451706290245, -0.014481737278401852, -0.030650142580270767, 0.07144446671009064, -0.10398940742015839, 0.008648036047816277, 0.12725894153118134, 0.032193928956985474, 0.12150469422340393, 0.05755417048931122, -0.06475795060396194, 0.00674046203494072, 0.05408455431461334, -0.11475005000829697, -0.13466225564479828, -0.033541448414325714, -0.017460739240050316, -0.1301354169845581, 0.029329262673854828, 0.09684194624423981, -0.054747361689805984, -0.011597746051847935, -0.014912877231836319, 0.019270379096269608, -0.008228112012147903, 0.1531243473291397, 0.029094357043504715, 0.04721291363239288, -0.07724244892597198, 0.14601807296276093, 0.06753411144018173, -0.11464221775531769, 0.08045824617147446, 0.10350644588470459, -0.0771358385682106, -0.012829023413360119, 0.07178323715925217, 0.1663295328617096, -0.04548769071698189, -0.05006902664899826, -0.0859282910823822, -0.08784393221139908, 0.05489552021026611, 0.07898757606744766, 0.03517286852002144, -0.005431924015283585, -0.04430869594216347, 0.05142498388886452, -0.1497107446193695, 0.08318644016981125, 0.05357448011636734, 0.06823580712080002, -0.15483805537223816, 0.09807517379522324, 0.01707305759191513, 0.028118999674916267, -0.01191699504852295, 0.0002184256591135636, -0.10452799499034882, -0.026387721300125122, -0.1225077286362648, -0.009003367274999619, -0.058412253856658936, 0.00721708731725812, -0.010395837016403675, -0.041929472237825394, -0.05294368416070938, 0.04898005351424217, -0.0671515017747879, -0.0699988454580307, 0.003913191147148609, 0.0633576288819313, -0.12473975121974945, 0.010649009607732296, 0.008004390634596348, -0.1043422743678093, 0.07003886997699738, 0.07544293999671936, 0.021540898829698563, 0.02830885350704193, -0.11754770576953888, -0.016038548201322556, 0.03095385432243347, 0.018380215391516685, 0.056321628391742706, -0.08880328387022018, -0.007003025617450476, -0.01574094593524933, 0.03811610862612724, -0.001581897959113121, 0.050038259476423264, -0.12935827672481537, -0.04254150390625, -0.06792767345905304, -0.05657881125807762, -0.07921294867992401, 0.05003219470381737, 0.1267199069261551, 0.020756417885422707, 0.16898733377456665, -0.07775361835956573, 0.04888836294412613, -0.17483952641487122, -0.01674710586667061, -0.0032267021015286446, -0.022682053968310356, -0.06887952983379364, -0.028893886134028435, 0.0784858763217926, -0.062287069857120514, 0.12199974805116653, -0.03499434515833855, 0.06907646358013153, 0.030823225155472755, -0.053794827312231064, -0.03362748399376869, 0.012370086275041103, 0.21867837011814117, 0.08822289109230042, -0.012078064493834972, 0.08345227688550949, -0.014588496647775173, 0.06458799540996552, 0.0365358330309391, 0.1776399165391922, 0.13365831971168518, -0.006219886243343353, 0.09938579052686691, 0.08194330334663391, -0.09592336416244507, -0.15187010169029236, 0.10312584042549133, -0.023440996184945107, 0.12209625542163849, -0.03914488106966019, 0.15846799314022064, 0.09792204946279526, -0.1491066813468933, 0.030582115054130554, -0.028875689953565598, -0.09419640153646469, -0.11894013732671738, -0.04336022585630417, -0.08429068326950073, -0.13400006294250488, 0.026010366156697273, -0.11606568843126297, 0.03157178312540054, 0.022633761167526245, 0.01782764494419098, 0.010889691300690174, 0.12524892389774323, 0.005525005515664816, 0.006265776231884956, 0.061027396470308304, 0.014449086040258408, -0.00808432511985302, -0.0464005284011364, -0.0851517915725708, 0.02727087214589119, 0.00043569892295636237, 0.07762160897254944, -0.03633089363574982, -0.021748367697000504, 0.05730613321065903, -0.014471317641437054, -0.07137658447027206, 0.03473377600312233, 0.004820255097001791, 0.026800120249390602, 0.046034637838602066, 0.04659554362297058, -0.02924368344247341, -0.05048530176281929, 0.24346953630447388, -0.07927645742893219, -0.05500631406903267, -0.11998743563890457, 0.18754541873931885, 0.03218702971935272, -0.01157071441411972, 0.06318649649620056, -0.10946249961853027, -0.043084293603897095, 0.1544322520494461, 0.14443114399909973, -0.03534526377916336, -0.02257664129137993, 0.010493390262126923, -0.01952536404132843, -0.05812042951583862, 0.1207042932510376, 0.10113391280174255, 0.06388405710458755, -0.04332060366868973, -0.013113177381455898, -0.0240167323499918, -0.033104658126831055, -0.09687425196170807, 0.060115981847047806, 0.01880836673080921, -0.012102290987968445, -0.03162316977977753, 0.0678781196475029, -0.03357364237308502, -0.1455298662185669, 0.024896902963519096, -0.13706928491592407, -0.1674313098192215, -0.030195388942956924, 0.10422209650278091, -0.012061214074492455, 0.05776016041636467, 0.006176827009767294, -0.0246463343501091, 0.12205470353364944, -0.006053504999727011, -0.07240858674049377, -0.08298038691282272, 0.06541427969932556, -0.05587451532483101, 0.2019103616476059, 0.008848896250128746, 0.0700521171092987, 0.11009620130062103, 0.05568929761648178, -0.13657328486442566, 0.02316007763147354, 0.0728272795677185, -0.08678669482469559, 0.04639347270131111, 0.15315891802310944, -0.05123777315020561, 0.08086852729320526, 0.04008067771792412, -0.08988158404827118, -0.01350813452154398, -0.06874246150255203, -0.013318520970642567, -0.06177090108394623, 0.001490390975959599, -0.060210052877664566, 0.15920284390449524, 0.19944319128990173, -0.031029589474201202, -0.016514338552951813, -0.0680902898311615, 0.03840731829404831, 0.043768804520368576, 0.09785381704568863, -0.026542088016867638, -0.21985091269016266, 0.009732979349792004, 0.032694049179553986, 0.04526698589324951, -0.21426817774772644, -0.09810946881771088, 0.035169679671525955, -0.04930848628282547, -0.07334353774785995, 0.09055111557245255, 0.06726758182048798, 0.024727411568164825, -0.04271971434354782, -0.10899978876113892, -0.04693683236837387, 0.13867925107479095, -0.15545232594013214, -0.040631189942359924 ]
null
null
transformers
#Ayjayo
{"tags": ["conversational"]}
text-generation
Ayjayo/DialoGPT-medium-AyjayoAI
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
#Ayjayo
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.009697278961539268, 0.03208012506365776, -0.007204889785498381, 0.004809224978089333, 0.16726240515708923, 0.014898733235895634, 0.09765533357858658, 0.13672804832458496, -0.007841327227652073, -0.031050153076648712, 0.14490588009357452, 0.20411323010921478, -0.006439372431486845, 0.0661218985915184, -0.07572533935308456, -0.2683109939098358, 0.05759621039032936, 0.046649303287267685, 0.016515716910362244, 0.1200079694390297, 0.08573378622531891, -0.05473608896136284, 0.08714032918214798, -0.014583407901227474, -0.150366872549057, 0.017733458429574966, 0.043394338339567184, -0.12260226160287857, 0.11910516023635864, 0.05462685227394104, 0.07063519209623337, 0.014929565601050854, -0.07541623711585999, -0.1631229966878891, 0.03031250834465027, 0.01425902172923088, -0.0594632662832737, 0.04757995903491974, 0.059961482882499695, -0.10165371745824814, 0.10819483548402786, 0.09530027210712433, -0.013078106567263603, 0.06798283755779266, -0.16849711537361145, -0.020869607105851173, -0.01446688175201416, 0.009899779222905636, 0.05550243332982063, 0.09964893013238907, -0.03413357585668564, 0.10497362166643143, -0.09214533120393753, 0.11017382889986038, 0.10932035744190216, -0.32057443261146545, -0.005767723545432091, 0.09167823940515518, 0.039358653128147125, 0.07352814823389053, -0.04467793554067612, 0.06258884817361832, 0.018015462905168533, 0.017986174672842026, -0.014015024527907372, -0.07283061742782593, -0.11612214148044586, 0.04717336222529411, -0.08668071031570435, -0.059868961572647095, 0.2244078367948532, -0.05464440956711769, 0.06881742179393768, -0.05281897634267807, -0.10522868484258652, -0.04308144748210907, -0.029833965003490448, 0.00475557055324316, -0.07660607248544693, 0.08692064881324768, 0.00869679357856512, -0.09547875821590424, -0.1376667022705078, -0.02496783249080181, -0.1776352822780609, 0.16140350699424744, 0.02465328387916088, 0.05232657864689827, -0.2027255892753601, 0.09623090922832489, 0.017906051129102707, -0.08045592904090881, 0.022091427817940712, -0.10046248883008957, 0.029131146147847176, 0.013760408386588097, -0.04754498973488808, -0.061387211084365845, 0.0843690037727356, 0.11199145019054413, -0.01731434464454651, 0.025486016646027565, -0.039331406354904175, 0.08100687712430954, 0.03553595021367073, 0.09077847748994827, 0.007288969587534666, -0.028338588774204254, 0.025842782109975815, -0.13719046115875244, -0.003647835226729512, -0.07116208970546722, -0.16572439670562744, -0.021088803187012672, 0.02994808368384838, 0.08289173990488052, 0.015449047088623047, 0.11682453751564026, -0.03272046521306038, -0.025152435526251793, 0.03602350503206253, -0.047656361013650894, -0.012649794109165668, 0.016648368909955025, 0.013163427822291851, 0.12399329990148544, -0.0022096503525972366, 0.03235051408410072, -0.13653022050857544, 0.031423524022102356, -0.06793295592069626, -0.003740974934771657, -0.03486552834510803, -0.040637075901031494, 0.009043924510478973, -0.06862333416938782, 0.003486064961180091, -0.15030112862586975, -0.15063877403736115, 0.007587034720927477, -0.007836631499230862, -0.04107699543237686, -0.06370922178030014, -0.06952770054340363, -0.013550350442528725, 0.04251532256603241, -0.07093454152345657, -0.011352915316820145, -0.06403283774852753, 0.11004766076803207, -0.03197755664587021, 0.07921615242958069, -0.11953279376029968, 0.08390819281339645, -0.11260783672332764, -0.02386913076043129, -0.060801517218351364, 0.09317506104707718, -0.0006014376995153725, 0.09549830108880997, -0.006563255097717047, -0.017931854352355003, -0.07981178909540176, 0.06445012241601944, -0.042872510850429535, 0.21701598167419434, -0.0615808479487896, -0.11181682348251343, 0.28781595826148987, -0.052628401666879654, -0.1370542049407959, 0.11647392809391022, 0.008682746440172195, 0.05777018144726753, 0.10703510791063309, 0.19733482599258423, -0.015276194550096989, 0.004040541127324104, 0.09471915662288666, 0.11263324320316315, -0.11276852339506149, -0.033160366117954254, 0.013019153848290443, -0.04081077128648758, -0.10867965966463089, 0.04689536616206169, 0.09810488671064377, 0.07090286910533905, -0.04786505550146103, -0.03377414867281914, -0.01366397924721241, 0.0052589005790650845, 0.08885077387094498, -0.007157256826758385, 0.10962837189435959, -0.05819983780384064, -0.03796621412038803, -0.029282379895448685, -0.012126247398555279, -0.03951939567923546, 0.03137664496898651, -0.043376367539167404, 0.10821941494941711, -0.011204327456653118, 0.06364280730485916, -0.16185984015464783, -0.07691477984189987, -0.017002692446112633, 0.1581239402294159, 0.024538565427064896, 0.09859629720449448, 0.0552486926317215, -0.040398042649030685, -0.0012767292791977525, 0.012792680412530899, 0.15581141412258148, -0.022091681137681007, -0.065607450902462, -0.052166227251291275, 0.08642971515655518, -0.05641226842999458, 0.04504093527793884, -0.05937713757157326, 0.012367865070700645, 0.05064384639263153, 0.10342344641685486, -0.00018274025933351368, 0.03323284164071083, -0.008164864964783192, 0.002145637758076191, -0.058205123990774155, 0.007405933458358049, 0.10799351334571838, 0.00036868182360194623, -0.07365862280130386, 0.22074243426322937, -0.17796069383621216, 0.1765957772731781, 0.1893044263124466, -0.299345999956131, 0.017949223518371582, -0.10759581625461578, -0.04561871662735939, 0.014407722279429436, 0.05567655712366104, -0.0454222597181797, 0.1703362911939621, -0.009871348738670349, 0.18874616920948029, -0.04946064203977585, -0.04464937001466751, -0.0200483538210392, -0.05118836089968681, -0.0024189651012420654, 0.07781197130680084, 0.10685696452856064, -0.13992026448249817, 0.1964332014322281, 0.1621224284172058, 0.048237916082143784, 0.19945049285888672, 0.015346456319093704, -0.011589210480451584, 0.0909530371427536, 0.005220826715230942, -0.058739423751831055, -0.07409929484128952, -0.2594851851463318, -0.030033592134714127, 0.07992640137672424, 0.0422382652759552, 0.1212305948138237, -0.11349532753229141, -0.038956157863140106, -0.01763172075152397, -0.023146281018853188, 0.021672505885362625, 0.0914369598031044, 0.06075398623943329, 0.13201528787612915, -0.001710098935291171, -0.007300339173525572, 0.10524573177099228, 0.01783694699406624, -0.09354141354560852, 0.18308524787425995, -0.13652534782886505, -0.37097251415252686, -0.13911493122577667, -0.18057456612586975, -0.05449081212282181, 0.05712554603815079, 0.11679314076900482, -0.12011238187551498, -0.018752124160528183, 0.01578843593597412, 0.10931742936372757, -0.08449502289295197, 0.0021454424131661654, -0.06880278885364532, 0.0321490578353405, -0.10310184955596924, -0.09194442629814148, -0.055416494607925415, -0.031392451375722885, -0.08001253753900528, 0.1423761546611786, -0.10777941346168518, 0.04476889222860336, 0.20262959599494934, 0.04653622955083847, 0.05625178664922714, -0.044105201959609985, 0.19377262890338898, -0.11264272034168243, -0.01661740615963936, 0.19215328991413116, -0.048360925167798996, 0.07476246356964111, 0.1232115849852562, -0.006348740309476852, -0.08765771239995956, 0.03011748194694519, -0.02085109055042267, -0.07988511025905609, -0.23219464719295502, -0.13938382267951965, -0.12429051846265793, 0.09477275609970093, 0.028005298227071762, 0.056365787982940674, 0.17219258844852448, 0.06577219814062119, -0.038416244089603424, 0.006410336587578058, 0.02959546446800232, 0.08237514644861221, 0.23417828977108002, -0.06035616248846054, 0.1364797055721283, -0.03420931473374367, -0.14982740581035614, 0.08169995993375778, 0.0713929831981659, 0.10213395953178406, 0.06678459793329239, 0.0804823637008667, 0.0149586396291852, 0.06188136339187622, 0.1311223804950714, 0.08191446959972382, 0.019586285576224327, -0.02480296604335308, -0.03388110175728798, -0.025523077696561813, -0.05937909707427025, 0.040128443390131, 0.06589099019765854, -0.16763372719287872, -0.039227183908224106, -0.09338314831256866, 0.09657008945941925, 0.0873042419552803, 0.06609832495450974, -0.1842060089111328, -0.008006223477423191, 0.08488986641168594, -0.03854905813932419, -0.13727426528930664, 0.09535189718008041, 0.01523482333868742, -0.15144726634025574, 0.03139317408204079, -0.04061909019947052, 0.12188644707202911, -0.07804752141237259, 0.09809603542089462, -0.08108244836330414, -0.07448557764291763, 0.02123199962079525, 0.1261177361011505, -0.30527687072753906, 0.20240111649036407, -0.0024993624538183212, -0.06486981362104416, -0.1243603527545929, -0.0032166161108762026, 0.002410882618278265, 0.07357452809810638, 0.10519039630889893, -0.007196315098553896, 0.001897757756523788, -0.06300821900367737, -0.01829923689365387, 0.032471053302288055, 0.13080233335494995, -0.0401318334043026, -0.021158374845981598, -0.050194524228572845, -0.001653497340157628, -0.03173094615340233, -0.06934895366430283, 0.02002747356891632, -0.19509181380271912, 0.08751901984214783, 0.04166261479258537, 0.09648149460554123, 0.029994789510965347, 0.004265148192644119, -0.09651939570903778, 0.24698667228221893, -0.07148019969463348, -0.10072879493236542, -0.10919588059186935, -0.046813901513814926, 0.03569883480668068, -0.05628936365246773, 0.04309194162487984, -0.0788632407784462, 0.028997479006648064, -0.06352769583463669, -0.19235502183437347, 0.12410202622413635, -0.09027006477117538, -0.04412810131907463, -0.02371402643620968, 0.2110891044139862, -0.05598580464720726, 0.010335659608244896, 0.02930437959730625, 0.01208863127976656, -0.11645778268575668, -0.09678568691015244, 0.031018631532788277, -0.007351789623498917, 0.050603240728378296, 0.041841957718133926, -0.05915454775094986, -0.017138581722974777, -0.052199993282556534, -0.022926922887563705, 0.3496883809566498, 0.14231905341148376, -0.043836336582899094, 0.19347235560417175, 0.12347975373268127, -0.07452994585037231, -0.3159443140029907, -0.1066238060593605, -0.10937739163637161, -0.04680149629712105, -0.07012093812227249, -0.2002030611038208, 0.06474938243627548, 0.00662544509395957, -0.013415241613984108, 0.12749312818050385, -0.2561831772327423, -0.07571036368608475, 0.15906259417533875, -0.017980827018618584, 0.3745945692062378, -0.1168576180934906, -0.10926306992769241, -0.03950892388820648, -0.14175476133823395, 0.16968177258968353, -0.01989765651524067, 0.11221715062856674, -0.009765521623194218, 0.14388824999332428, 0.05548359826207161, -0.023479344323277473, 0.08544106781482697, 0.004999885335564613, -0.03290518373250961, -0.10304180532693863, -0.05676887184381485, 0.007092386484146118, 0.02477436140179634, 0.018026655539870262, -0.041834570467472076, 0.02227151393890381, -0.11731979995965958, -0.04657655209302902, -0.08982590585947037, 0.04431166127324104, 0.03899754583835602, -0.07325074821710587, -0.002380647463724017, -0.07165111601352692, -0.012272949330508709, 0.022334342822432518, 0.20356793701648712, -0.08029330521821976, 0.16448934376239777, 0.09239562600851059, 0.12419285625219345, -0.14376309514045715, -0.00019283240544609725, -0.0762530043721199, -0.05611240118741989, 0.07737895101308823, -0.09433035552501678, 0.058893077075481415, 0.10901971161365509, -0.04567738622426987, 0.08828683942556381, 0.10377411544322968, 0.008936077356338501, 0.003213887568563223, 0.10916902124881744, -0.2667325437068939, -0.0296600554138422, -0.07532413303852081, 0.000883326749317348, 0.09092561900615692, 0.08562852442264557, 0.18840822577476501, 0.025361526757478714, -0.04293036088347435, -0.002770674182102084, 0.028597986325621605, -0.039021048694849014, 0.051667019724845886, 0.001123449532315135, 0.01947369985282421, -0.1530752182006836, 0.072522833943367, 0.01490565575659275, -0.15215420722961426, 0.021316176280379295, 0.16572684049606323, -0.11656328290700912, -0.1283872276544571, -0.06520111113786697, 0.08313824236392975, -0.11755692958831787, -0.01578943058848381, -0.03279297426342964, -0.13145680725574493, 0.07992171496152878, 0.12629036605358124, 0.05557859688997269, 0.0972496047616005, -0.06061713397502899, -0.020469192415475845, -0.018721895292401314, -0.014099318534135818, -0.012384648434817791, -0.007667020428925753, -0.055978111922740936, 0.0590752474963665, -0.026677248999476433, 0.1425808072090149, -0.09221141785383224, -0.1037059873342514, -0.16142144799232483, 0.0374140702188015, -0.11013076454401016, -0.08825794607400894, -0.08821134269237518, -0.050188567489385605, 0.002360827289521694, -0.019856395199894905, -0.04037635400891304, -0.05829505994915962, -0.12300454825162888, 0.0338277705013752, -0.040771447122097015, 0.024727050215005875, -0.07512269169092178, 0.015856385231018066, 0.08507686108350754, -0.03285100311040878, 0.15655414760112762, 0.1450488418340683, -0.1006515845656395, 0.10741901397705078, -0.14806775748729706, -0.09138492494821548, 0.11116421222686768, 0.015329592861235142, 0.0449691042304039, 0.09723787009716034, 0.013362943194806576, 0.0635865181684494, 0.032776717096567154, 0.05308786407113075, 0.027619892731308937, -0.11959987878799438, 0.06483134627342224, -0.03626115620136261, -0.14700546860694885, -0.049338050186634064, -0.05282869189977646, 0.01647452637553215, 0.013054544106125832, 0.09622690081596375, -0.05301849544048309, 0.10698331147432327, -0.04055701196193695, 0.0346808135509491, 0.017554637044668198, -0.1730053424835205, -0.03816922754049301, -0.08538098633289337, 0.03681723028421402, 0.014741539023816586, 0.25266793370246887, 0.030072299763560295, 0.012416383251547813, 0.032671261578798294, 0.08285367488861084, 0.03899408504366875, 0.010228337720036507, 0.17482228577136993, 0.1162426546216011, -0.06621865928173065, -0.10445023328065872, 0.0729617029428482, 0.016332454979419708, 0.01286179106682539, 0.13617953658103943, 0.008365051820874214, 0.005795429926365614, 0.08649782836437225, -0.016865963116288185, 0.009968153201043606, -0.10052056610584259, -0.13426925241947174, -0.022176474332809448, 0.05151832848787308, -0.04655967652797699, 0.11727844923734665, 0.1406494379043579, -0.01806013658642769, 0.03222079202532768, -0.021771740168333054, -0.05699979141354561, -0.1683429479598999, -0.1429590880870819, -0.06883849948644638, -0.13416796922683716, 0.00897989235818386, -0.11180389672517776, 0.05395037308335304, 0.06001098081469536, 0.06750501692295074, -0.06899319589138031, 0.10220931470394135, 0.04626858979463577, -0.11440542340278625, 0.06264589726924896, -0.0296088308095932, 0.09430401772260666, -0.02759445086121559, -0.019505485892295837, -0.09039592742919922, 0.014574515633285046, 0.011419114656746387, 0.06245238706469536, -0.04707273095846176, 0.007463190704584122, -0.14696238934993744, -0.08972041308879852, -0.0523175448179245, 0.0718572810292244, -0.050409089773893356, 0.14282815158367157, 0.00775480642914772, -0.0170906875282526, 0.039554283022880554, 0.22787313163280487, -0.07476283609867096, -0.04778539761900902, -0.05269690603017807, 0.20717895030975342, 0.02975541539490223, 0.1171872541308403, -0.022938819602131844, -0.006106364540755749, -0.0919521227478981, 0.3764844834804535, 0.30030161142349243, -0.09031439572572708, 0.011794124729931355, 0.02137952297925949, 0.04502861574292183, 0.1316293478012085, 0.1216534823179245, 0.10318691283464432, 0.3006802201271057, -0.07452366501092911, -0.04653361067175865, -0.012629742734134197, -0.023858042433857918, -0.09059546142816544, 0.1021224707365036, 0.04839762672781944, -0.06382183730602264, -0.03313443064689636, 0.0954432487487793, -0.25862133502960205, 0.1277991235256195, -0.12311873584985733, -0.17578600347042084, -0.06654827296733856, 0.009760108776390553, 0.10465722531080246, 0.015642458572983742, 0.0946015790104866, 0.007128213066607714, -0.11252258718013763, 0.06305865943431854, 0.03397420793771744, -0.22762253880500793, 0.0006893770187161863, 0.06642123311758041, -0.07006710022687912, -0.0024247700348496437, -0.026499588042497635, 0.05657242611050606, 0.0656052976846695, 0.054629553109407425, -0.00971333310008049, 0.03816632181406021, 0.0034184439573436975, -0.0585215799510479, 0.016623929142951965, 0.05121519789099693, 0.02472509816288948, -0.09763528406620026, 0.06927435845136642, -0.1574270874261856, 0.04766253009438515, -0.0030655991286039352, -0.04124255105853081, 0.006064958870410919, 0.008823691867291927, -0.06491616368293762, 0.05165379121899605, 0.07916834205389023, -0.0016257909592241049, -0.0062433634884655476, -0.057178743183612823, -0.02632102556526661, -0.027755750343203545, -0.09291748702526093, -0.10495562851428986, -0.14682936668395996, -0.11640441417694092, 0.09368976950645447, -0.01011267676949501, -0.1848134547472, 0.022154374048113823, -0.08606051653623581, 0.08319322764873505, -0.1670055389404297, 0.08040720224380493, 0.07041648775339127, 0.013038921169936657, -0.0031511052511632442, -0.02002427540719509, 0.054132770746946335, 0.086809903383255, -0.10407156497240067, -0.07400695979595184 ]
null
null
transformers
在2.5亿的中文语料上,进行mobie_bert进行预训练。在单卡-A100下迭代100万 steps,训练15天。
{"license": "apache-2.0"}
fill-mask
Ayou/chinese_mobile_bert
[ "transformers", "pytorch", "mobilebert", "fill-mask", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #mobilebert #fill-mask #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
在2.5亿的中文语料上,进行mobie_bert进行预训练。在单卡-A100下迭代100万 steps,训练15天。
[]
[ "TAGS\n#transformers #pytorch #mobilebert #fill-mask #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 45 ]
[ "passage: TAGS\n#transformers #pytorch #mobilebert #fill-mask #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.039620291441679, 0.05711764842271805, -0.006843444425612688, 0.006029057316482067, 0.09265536069869995, -0.00574586121365428, 0.13673005998134613, 0.08963460475206375, -0.003938381560146809, -0.061505015939474106, 0.162319615483284, 0.20944635570049286, 0.0029658894054591656, 0.09806286543607712, -0.0421922504901886, -0.18433260917663574, 0.06630182266235352, 0.022128604352474213, -0.059943314641714096, 0.12049916386604309, 0.09122274070978165, -0.01377980224788189, 0.07665444165468216, 0.008035428822040558, -0.06684491783380508, 0.03938138484954834, 0.060861773788928986, -0.08274459838867188, 0.0973159447312355, -0.018364233896136284, 0.11158062517642975, 0.02840096689760685, -0.030616765841841698, -0.1341572105884552, 0.02318454720079899, 0.019440853968262672, -0.062223322689533234, 0.030605440959334373, 0.0035100809764117002, -0.04141465574502945, 0.01772349141538143, 0.04522913694381714, 0.01004329789429903, 0.05615298077464104, -0.06120247766375542, -0.20007790625095367, -0.10014857351779938, 0.0636407807469368, 0.058582719415426254, 0.08553081005811691, 0.05956842005252838, 0.20581106841564178, -0.10342329740524292, 0.07111424207687378, 0.16292105615139008, -0.3550746738910675, 0.0029767018277198076, 0.0611884742975235, 0.08118680119514465, -0.0778869017958641, 0.01944880373775959, 0.03535037487745285, 0.04264901950955391, 0.04569219425320625, 0.05152067914605141, -0.05625790357589722, -0.07299439609050751, 0.0013454982545226812, -0.04112819582223892, -0.07637084275484085, 0.1912619173526764, -0.026511764153838158, 0.025532621890306473, -0.007097487337887287, -0.10610752552747726, 0.0351392924785614, -0.018137866631150246, 0.022550417110323906, 0.021964650601148605, 0.08500166237354279, 0.01944386400282383, -0.03538213297724724, -0.11507754772901535, 0.013390087522566319, -0.2459254413843155, 0.10630296170711517, 0.029975224286317825, 0.08183740824460983, -0.16321535408496857, 0.05890994146466255, 0.04395582899451256, -0.1364835500717163, 0.028227584436535835, -0.0705387070775032, 0.12475352734327316, 0.03605607897043228, -0.03298189118504524, 0.010689589194953442, 0.1296139508485794, 0.22332851588726044, 0.08614743500947952, 0.009187816642224789, -0.03752756491303444, 0.09048109501600266, -0.04030897095799446, 0.037838153541088104, 0.04704160615801811, -0.07089473307132721, 0.09689219295978546, -0.11991024762392044, 0.10528291761875153, -0.023937147110700607, -0.1287282258272171, -0.021342379972338676, -0.05113577842712402, 0.13435761630535126, 0.08761775493621826, 0.023231683298945427, -0.060688622295856476, 0.015686750411987305, 0.18439501523971558, -0.04667576029896736, -0.008588948287069798, -0.030995314940810204, 0.05167315527796745, 0.07093130797147751, 0.06664019823074341, 0.011860299855470657, -0.025000188499689102, 0.05068085715174675, -0.04662618786096573, -0.057109035551548004, -0.038078922778367996, 0.00285727484151721, 0.07438518106937408, -0.09585598856210709, 0.02546824887394905, -0.1809963434934616, -0.19305546581745148, 0.07473745942115784, 0.07542083412408829, 0.012529078871011734, -0.08435465395450592, 0.05655280873179436, -0.009800422936677933, 0.014158356934785843, -0.07394950091838837, 0.00014288023521658033, -0.05131128057837486, 0.08361845463514328, -0.04052632302045822, 0.04508913680911064, -0.11970199644565582, 0.04817229509353638, -0.10565947741270065, 0.025322435423731804, -0.05769572779536247, -0.07246100902557373, -0.07284630089998245, 0.18061091005802155, -0.0449562668800354, -0.03276529163122177, -0.032697807997465134, 0.00046879355795681477, -0.03049299865961075, 0.11574026197195053, -0.034409187734127045, -0.11008058488368988, 0.1733635514974594, -0.12535303831100464, -0.1967330127954483, 0.05033092945814133, 0.03009776771068573, 0.025146152824163437, 0.041881151497364044, 0.09510910511016846, 0.06823961436748505, -0.13921713829040527, 0.06246679648756981, 0.1193738728761673, -0.14573729038238525, -0.21875643730163574, 0.03870546817779541, -0.0585351325571537, -0.15611431002616882, 0.06088234484195709, 0.025398865342140198, 0.14920440316200256, -0.013268153183162212, -0.0774586945772171, -0.05148758366703987, -0.0737619698047638, 0.04938273876905441, 0.005913734436035156, 0.05283345654606819, -0.07919762283563614, 0.0018575022695586085, -0.09624602645635605, 0.04276128113269806, 0.08048488199710846, 0.03929254785180092, -0.09722280502319336, 0.09665807336568832, -0.001041888608597219, 0.004793537314981222, -0.08179312944412231, -0.058070797473192215, 0.01293859165161848, -0.046450790017843246, -0.03628513216972351, 0.043424610048532486, 0.08681134879589081, -0.04739616438746452, -0.0127863809466362, -0.01143793947994709, 0.10115519165992737, 0.06278173625469208, 0.0030071602668613195, -0.1193334311246872, 0.034273911267519, -0.03698912262916565, 0.02134418673813343, 0.052309826016426086, 0.02294532209634781, 0.005963626783341169, 0.1346575766801834, -0.03232301026582718, 0.061112068593502045, -0.06324541568756104, -0.027223357930779457, -0.03647667169570923, -0.02255988121032715, 0.10430778563022614, 0.05026492848992348, -0.00236586038954556, 0.2097214311361313, -0.07364096492528915, 0.29866933822631836, 0.19009409844875336, -0.17794080078601837, 0.034858882427215576, 0.06506380438804626, -0.033336445689201355, -0.04397020861506462, 0.03307154029607773, 0.014144591987133026, -0.011911780573427677, 0.018131714314222336, 0.16876423358917236, -0.024555101990699768, -0.027612455189228058, 0.011757670901715755, -0.061552561819553375, -0.003708191681653261, 0.006583277136087418, 0.17343474924564362, -0.12101679295301437, 0.16427171230316162, 0.34129926562309265, 0.0021889102645218372, 0.08028952777385712, -0.049205418676137924, 0.03607204928994179, -0.013073036447167397, -0.047454580664634705, -0.01748005673289299, 0.058110836893320084, -0.13876421749591827, 0.016894280910491943, 0.07270568609237671, -0.011155621148645878, 0.07413023710250854, -0.09359847009181976, -0.04927261546254158, 0.027400610968470573, 0.020683525130152702, -0.06453680992126465, 0.12499940395355225, 0.003759522922337055, 0.04884595423936844, -0.04627946391701698, -0.12646949291229248, 0.13940049707889557, 0.005723216105252504, -0.0418621227145195, 0.11167839914560318, -0.1712670624256134, -0.29253077507019043, -0.11301761120557785, -0.19599640369415283, -0.03158722072839737, 0.004753932822495699, 0.11046093702316284, -0.04611573368310928, -0.043129995465278625, 0.04775947332382202, -0.07724922895431519, 0.0407528355717659, 0.026900600641965866, -0.036771390587091446, 0.026326598599553108, 0.031315118074417114, -0.12032657116651535, -0.06249194219708443, 0.038810886442661285, -0.06761184334754944, 0.078217513859272, -0.09386446326971054, 0.06723306328058243, 0.10074355453252792, 0.03071453794836998, 0.03735163435339928, -0.003125061746686697, 0.11755649745464325, -0.02809961512684822, 0.004160887096077204, 0.23473496735095978, 0.0644599124789238, 0.10045990347862244, 0.1405014544725418, 0.01673828437924385, -0.03223530203104019, 0.003363721538335085, -0.04187425971031189, -0.09437579661607742, -0.23039154708385468, -0.1193779930472374, -0.09458482265472412, 0.01592220552265644, 0.038752440363168716, 0.0942283347249031, 0.17188677191734314, 0.08461914211511612, 0.023183053359389305, -0.012383640743792057, -0.04976539686322212, 0.05577448010444641, 0.20047080516815186, -0.03611692041158676, 0.09868139773607254, -0.09998428821563721, -0.0838366225361824, 0.1143852025270462, 0.06475180387496948, 0.09280850738286972, 0.11751548945903778, 0.003674386767670512, 0.0717654600739479, 0.23105832934379578, 0.06450188905000687, 0.1268509477376938, 0.035010695457458496, -0.03682275861501694, -0.033537574112415314, -0.011000090278685093, -0.04470068961381912, 0.009290044195950031, 0.08350007236003876, -0.11370699852705002, -0.025054898113012314, -0.12027647346258163, 0.05837877467274666, 0.20990090072155, 0.020002348348498344, -0.14606539905071259, 0.06934984028339386, 0.0657307356595993, -0.028819289058446884, -0.03579740971326828, 0.08659036457538605, -0.051324985921382904, -0.1291833072900772, 0.1308954507112503, -0.04288920760154724, 0.09546775370836258, 0.06783034652471542, 0.054586950689554214, 0.017110800370573997, -0.0797555223107338, 0.07976407557725906, 0.1210865005850792, -0.25141024589538574, 0.20715560019016266, -0.029419826343655586, -0.03183780238032341, -0.08259107917547226, 0.008741188794374466, 0.07735593616962433, 0.17172889411449432, 0.08481484651565552, 0.06063290685415268, -0.016316769644618034, -0.006891550961881876, -0.06287211924791336, 0.04456565901637077, 0.0054456028155982494, -0.008745010942220688, -0.0628431960940361, -0.054281122982501984, -0.054014746099710464, -0.003951631486415863, 0.042178262025117874, -0.05815109983086586, -0.11747504025697708, 0.06357857584953308, 0.1040610671043396, 0.009915498085319996, -0.08573666214942932, -0.02383051998913288, -0.09985876828432083, 0.1913379430770874, -0.07585133612155914, -0.04565199464559555, -0.07689877599477768, -0.21583537757396698, 0.04497140273451805, -0.05314258486032486, 0.0944020003080368, -0.07839597761631012, -0.0053267572075128555, -0.0947493240237236, -0.1757841557264328, 0.12222308665513992, -0.1560615748167038, -0.04016043618321419, -0.06589455157518387, 0.1432231068611145, -0.089299775660038, 0.010710927657783031, 0.02314653992652893, -0.015665864571928978, -0.11266005039215088, -0.09166628867387772, -0.006823156028985977, -0.019527720287442207, 0.06975360959768295, 0.016431383788585663, -0.031027935445308685, -0.03133431822061539, 0.052038274705410004, -0.018059784546494484, 0.19539307057857513, 0.21818721294403076, -0.05374259501695633, 0.1746395230293274, 0.17954643070697784, -0.06606824696063995, -0.30528026819229126, -0.2005782276391983, -0.1438947170972824, -0.07304982095956802, 0.0452653132379055, -0.12456502765417099, 0.1001410186290741, 0.023876311257481575, -0.10007047653198242, 0.09082107245922089, -0.18494364619255066, -0.07734296470880508, 0.2143038660287857, -0.03452172130346298, 0.3850305378437042, -0.09543205797672272, -0.0678795874118805, -0.09961821883916855, -0.16105054318904877, 0.0863063856959343, -0.10700345784425735, 0.07048095762729645, -0.016833070665597916, 0.03929204121232033, -0.029199402779340744, -0.0872160866856575, 0.13078399002552032, -0.025372928008437157, 0.025177622213959694, -0.1031302809715271, -0.01714073307812214, 0.14300169050693512, -0.015768712386488914, 0.032188527286052704, -0.13047215342521667, 0.03424030542373657, -0.07542756199836731, 0.0070935687981545925, -0.09787304699420929, 0.1386946439743042, 0.008498825132846832, -0.04345176741480827, -0.01365432795137167, -0.033331677317619324, 0.04470420256257057, -0.008024957962334156, 0.15850132703781128, 0.05678364634513855, 0.062346525490283966, 0.15492196381092072, 0.011570111848413944, -0.16585083305835724, -0.06587962061166763, -0.1175851821899414, -0.0872160941362381, 0.05270695313811302, -0.04267903417348862, 0.04632220417261124, 0.07176575809717178, -0.04497889056801796, 0.03584592789411545, 0.06604928523302078, 0.005417767912149429, -0.055191099643707275, 0.1345401555299759, -0.16275060176849365, 0.05714384466409683, -0.0034385300241410732, 0.16567613184452057, 0.07098910957574844, 0.0653185322880745, 0.09731487184762955, 0.020789919421076775, -0.040349800139665604, 0.020875638350844383, 0.053171683102846146, -0.11089308559894562, 0.04003983363509178, 0.04724135249853134, 0.030836820602416992, -0.13404269516468048, 0.07420841604471207, -0.03286478668451309, -0.13366401195526123, -0.029155375435948372, 0.06761883944272995, -0.11915436387062073, -0.10642945766448975, 0.062204379588365555, 0.08994518220424652, -0.08366898447275162, -0.10629348456859589, -0.03789554908871651, -0.1291411817073822, 0.054894935339689255, 0.14455874264240265, 0.11693497747182846, 0.09942737966775894, -0.007491499651223421, -0.06063683331012726, 0.009636269882321358, -0.02068806253373623, -0.0345597006380558, 0.02732846513390541, -0.10054179280996323, -0.07799208909273148, 0.03809693455696106, 0.11805320531129837, -0.07607988268136978, -0.013830837793648243, -0.06638278812170029, 0.023820720613002777, -0.13990509510040283, -0.0636010691523552, -0.08590257167816162, -0.03856305405497551, 0.03910871595144272, -0.07038706541061401, -0.04168827831745148, -0.004194226115942001, -0.11539632081985474, -0.007761112879961729, 0.03990951180458069, 0.02134632132947445, -0.11219583451747894, -0.040481191128492355, 0.1197444275021553, -0.03374670073390007, 0.08814199268817902, 0.10692863166332245, -0.05866149812936783, 0.06678177416324615, -0.08982735872268677, -0.15779553353786469, 0.06723184138536453, 0.05289840325713158, 0.022292684763669968, -0.03105093538761139, 0.0008514939108863473, 0.11788371950387955, -0.007616611197590828, 0.03528532758355141, 0.07776028662919998, -0.14307208359241486, 0.020801689475774765, 0.00099231640342623, -0.14700162410736084, 0.002571034710854292, -0.12460270524024963, 0.09241852164268494, 0.00039017308154143393, 0.19635722041130066, -0.030136315152049065, 0.03331298381090164, -0.0351380929350853, 0.025095071643590927, -0.05626195669174194, -0.1528148502111435, -0.07668840140104294, -0.013732289895415306, -0.05532548949122429, -0.0030789636075496674, 0.2631887197494507, -0.019064530730247498, -0.07155834883451462, 0.0518258735537529, 0.08269725739955902, -0.009831523522734642, -0.01630699448287487, 0.16143788397312164, 0.051950275897979736, 0.0020715438295155764, -0.06530053913593292, 0.019925693050026894, 0.03208183869719505, -0.13620120286941528, 0.07616157084703445, 0.06671962887048721, 0.09560716152191162, 0.06138024851679802, 0.05680135264992714, -0.005872719921171665, -0.13865917921066284, -0.1784014254808426, -0.0034162572119385004, 0.08360841125249863, 0.01381616946309805, 0.047845508903265, 0.11283190548419952, -0.00908998679369688, 0.013566124252974987, -0.012273170053958893, 0.011000669561326504, -0.17263971269130707, -0.10375934094190598, -0.059142354875802994, -0.12700772285461426, 0.01485460251569748, 0.007498122751712799, 0.0004284705501049757, 0.11569692939519882, 0.03633297607302666, -0.03629054129123688, 0.03713929280638695, -0.10572312027215958, 0.009098697453737259, 0.00637782271951437, -0.017415208742022514, -0.020595496520400047, -0.027414187788963318, -0.041070107370615005, -0.15022413432598114, 0.010165422223508358, -0.03543975576758385, 0.02923794835805893, -0.031240195035934448, 0.060713991522789, -0.07996579259634018, -0.08183174580335617, -0.08426787704229355, 0.01337786391377449, -0.03085505962371826, 0.13411110639572144, 0.008896037936210632, 0.06240161508321762, 0.06801310181617737, 0.10270001739263535, -0.08882977813482285, -0.1852247714996338, -0.05433225631713867, 0.11870326101779938, 0.039167534559965134, 0.0613853894174099, 0.00041542560211382806, 0.06601271778345108, -0.11189540475606918, 0.3157089650630951, 0.30200374126434326, -0.03921577334403992, 0.0532451793551445, 0.008977389894425869, 0.022374257445335388, 0.03132840245962143, 0.08135586977005005, 0.10481731593608856, 0.22335770726203918, -0.08117010444402695, -0.06122146174311638, -0.04743995517492294, -0.03307158872485161, -0.15285764634609222, 0.025853559374809265, -0.04091668128967285, -0.08227342367172241, -0.0017886116402223706, 0.06421513110399246, -0.0857471451163292, 0.142672598361969, 0.045232538133859634, -0.10029052197933197, -0.037787750363349915, 0.006650255527347326, 0.2142564058303833, 0.042884618043899536, 0.04527553170919418, -0.02565007098019123, -0.051454611122608185, 0.14501194655895233, -0.004828573204576969, -0.22117577493190765, -0.09577598422765732, 0.11483220010995865, -0.030650557950139046, 0.18569058179855347, -0.012286441400647163, 0.04083952680230141, 0.0821324959397316, 0.05259009823203087, -0.07180958241224289, 0.05704108998179436, 0.019646331667900085, -0.09264280647039413, -0.03060302697122097, -0.11513945460319519, -0.02988469786942005, -0.08397458493709564, -0.015236059203743935, -0.05884551629424095, 0.031019918620586395, -0.0064727445133030415, -0.01922452636063099, -0.031328234821558, 0.053066425025463104, -0.04557375609874725, 0.05006859079003334, -0.001679536304436624, 0.0012639204505831003, -0.04186605289578438, -0.06187008321285248, -0.004892859607934952, 0.047263070940971375, -0.15605677664279938, -0.09576960653066635, -0.025346066802740097, -0.03504286706447601, 0.011379821226000786, 0.019093450158834457, -0.056142281740903854, -0.06328044086694717, -0.11202506721019745, -0.005975639447569847, -0.1620016247034073, 0.016290100291371346, 0.0627785250544548, 0.021736405789852142, 0.010541776195168495, 0.028660686686635017, 0.0022505633533000946, 0.03141801059246063, -0.1291068196296692, -0.09801243990659714 ]
null
null
transformers
#DialoGPT medium model (Harry Potter 1-3)
{"tags": ["conversational"]}
text-generation
Ayran/DialoGPT-medium-harry-potter-1-through-3
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
#DialoGPT medium model (Harry Potter 1-3)
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.009697278961539268, 0.03208012506365776, -0.007204889785498381, 0.004809224978089333, 0.16726240515708923, 0.014898733235895634, 0.09765533357858658, 0.13672804832458496, -0.007841327227652073, -0.031050153076648712, 0.14490588009357452, 0.20411323010921478, -0.006439372431486845, 0.0661218985915184, -0.07572533935308456, -0.2683109939098358, 0.05759621039032936, 0.046649303287267685, 0.016515716910362244, 0.1200079694390297, 0.08573378622531891, -0.05473608896136284, 0.08714032918214798, -0.014583407901227474, -0.150366872549057, 0.017733458429574966, 0.043394338339567184, -0.12260226160287857, 0.11910516023635864, 0.05462685227394104, 0.07063519209623337, 0.014929565601050854, -0.07541623711585999, -0.1631229966878891, 0.03031250834465027, 0.01425902172923088, -0.0594632662832737, 0.04757995903491974, 0.059961482882499695, -0.10165371745824814, 0.10819483548402786, 0.09530027210712433, -0.013078106567263603, 0.06798283755779266, -0.16849711537361145, -0.020869607105851173, -0.01446688175201416, 0.009899779222905636, 0.05550243332982063, 0.09964893013238907, -0.03413357585668564, 0.10497362166643143, -0.09214533120393753, 0.11017382889986038, 0.10932035744190216, -0.32057443261146545, -0.005767723545432091, 0.09167823940515518, 0.039358653128147125, 0.07352814823389053, -0.04467793554067612, 0.06258884817361832, 0.018015462905168533, 0.017986174672842026, -0.014015024527907372, -0.07283061742782593, -0.11612214148044586, 0.04717336222529411, -0.08668071031570435, -0.059868961572647095, 0.2244078367948532, -0.05464440956711769, 0.06881742179393768, -0.05281897634267807, -0.10522868484258652, -0.04308144748210907, -0.029833965003490448, 0.00475557055324316, -0.07660607248544693, 0.08692064881324768, 0.00869679357856512, -0.09547875821590424, -0.1376667022705078, -0.02496783249080181, -0.1776352822780609, 0.16140350699424744, 0.02465328387916088, 0.05232657864689827, -0.2027255892753601, 0.09623090922832489, 0.017906051129102707, -0.08045592904090881, 0.022091427817940712, -0.10046248883008957, 0.029131146147847176, 0.013760408386588097, -0.04754498973488808, -0.061387211084365845, 0.0843690037727356, 0.11199145019054413, -0.01731434464454651, 0.025486016646027565, -0.039331406354904175, 0.08100687712430954, 0.03553595021367073, 0.09077847748994827, 0.007288969587534666, -0.028338588774204254, 0.025842782109975815, -0.13719046115875244, -0.003647835226729512, -0.07116208970546722, -0.16572439670562744, -0.021088803187012672, 0.02994808368384838, 0.08289173990488052, 0.015449047088623047, 0.11682453751564026, -0.03272046521306038, -0.025152435526251793, 0.03602350503206253, -0.047656361013650894, -0.012649794109165668, 0.016648368909955025, 0.013163427822291851, 0.12399329990148544, -0.0022096503525972366, 0.03235051408410072, -0.13653022050857544, 0.031423524022102356, -0.06793295592069626, -0.003740974934771657, -0.03486552834510803, -0.040637075901031494, 0.009043924510478973, -0.06862333416938782, 0.003486064961180091, -0.15030112862586975, -0.15063877403736115, 0.007587034720927477, -0.007836631499230862, -0.04107699543237686, -0.06370922178030014, -0.06952770054340363, -0.013550350442528725, 0.04251532256603241, -0.07093454152345657, -0.011352915316820145, -0.06403283774852753, 0.11004766076803207, -0.03197755664587021, 0.07921615242958069, -0.11953279376029968, 0.08390819281339645, -0.11260783672332764, -0.02386913076043129, -0.060801517218351364, 0.09317506104707718, -0.0006014376995153725, 0.09549830108880997, -0.006563255097717047, -0.017931854352355003, -0.07981178909540176, 0.06445012241601944, -0.042872510850429535, 0.21701598167419434, -0.0615808479487896, -0.11181682348251343, 0.28781595826148987, -0.052628401666879654, -0.1370542049407959, 0.11647392809391022, 0.008682746440172195, 0.05777018144726753, 0.10703510791063309, 0.19733482599258423, -0.015276194550096989, 0.004040541127324104, 0.09471915662288666, 0.11263324320316315, -0.11276852339506149, -0.033160366117954254, 0.013019153848290443, -0.04081077128648758, -0.10867965966463089, 0.04689536616206169, 0.09810488671064377, 0.07090286910533905, -0.04786505550146103, -0.03377414867281914, -0.01366397924721241, 0.0052589005790650845, 0.08885077387094498, -0.007157256826758385, 0.10962837189435959, -0.05819983780384064, -0.03796621412038803, -0.029282379895448685, -0.012126247398555279, -0.03951939567923546, 0.03137664496898651, -0.043376367539167404, 0.10821941494941711, -0.011204327456653118, 0.06364280730485916, -0.16185984015464783, -0.07691477984189987, -0.017002692446112633, 0.1581239402294159, 0.024538565427064896, 0.09859629720449448, 0.0552486926317215, -0.040398042649030685, -0.0012767292791977525, 0.012792680412530899, 0.15581141412258148, -0.022091681137681007, -0.065607450902462, -0.052166227251291275, 0.08642971515655518, -0.05641226842999458, 0.04504093527793884, -0.05937713757157326, 0.012367865070700645, 0.05064384639263153, 0.10342344641685486, -0.00018274025933351368, 0.03323284164071083, -0.008164864964783192, 0.002145637758076191, -0.058205123990774155, 0.007405933458358049, 0.10799351334571838, 0.00036868182360194623, -0.07365862280130386, 0.22074243426322937, -0.17796069383621216, 0.1765957772731781, 0.1893044263124466, -0.299345999956131, 0.017949223518371582, -0.10759581625461578, -0.04561871662735939, 0.014407722279429436, 0.05567655712366104, -0.0454222597181797, 0.1703362911939621, -0.009871348738670349, 0.18874616920948029, -0.04946064203977585, -0.04464937001466751, -0.0200483538210392, -0.05118836089968681, -0.0024189651012420654, 0.07781197130680084, 0.10685696452856064, -0.13992026448249817, 0.1964332014322281, 0.1621224284172058, 0.048237916082143784, 0.19945049285888672, 0.015346456319093704, -0.011589210480451584, 0.0909530371427536, 0.005220826715230942, -0.058739423751831055, -0.07409929484128952, -0.2594851851463318, -0.030033592134714127, 0.07992640137672424, 0.0422382652759552, 0.1212305948138237, -0.11349532753229141, -0.038956157863140106, -0.01763172075152397, -0.023146281018853188, 0.021672505885362625, 0.0914369598031044, 0.06075398623943329, 0.13201528787612915, -0.001710098935291171, -0.007300339173525572, 0.10524573177099228, 0.01783694699406624, -0.09354141354560852, 0.18308524787425995, -0.13652534782886505, -0.37097251415252686, -0.13911493122577667, -0.18057456612586975, -0.05449081212282181, 0.05712554603815079, 0.11679314076900482, -0.12011238187551498, -0.018752124160528183, 0.01578843593597412, 0.10931742936372757, -0.08449502289295197, 0.0021454424131661654, -0.06880278885364532, 0.0321490578353405, -0.10310184955596924, -0.09194442629814148, -0.055416494607925415, -0.031392451375722885, -0.08001253753900528, 0.1423761546611786, -0.10777941346168518, 0.04476889222860336, 0.20262959599494934, 0.04653622955083847, 0.05625178664922714, -0.044105201959609985, 0.19377262890338898, -0.11264272034168243, -0.01661740615963936, 0.19215328991413116, -0.048360925167798996, 0.07476246356964111, 0.1232115849852562, -0.006348740309476852, -0.08765771239995956, 0.03011748194694519, -0.02085109055042267, -0.07988511025905609, -0.23219464719295502, -0.13938382267951965, -0.12429051846265793, 0.09477275609970093, 0.028005298227071762, 0.056365787982940674, 0.17219258844852448, 0.06577219814062119, -0.038416244089603424, 0.006410336587578058, 0.02959546446800232, 0.08237514644861221, 0.23417828977108002, -0.06035616248846054, 0.1364797055721283, -0.03420931473374367, -0.14982740581035614, 0.08169995993375778, 0.0713929831981659, 0.10213395953178406, 0.06678459793329239, 0.0804823637008667, 0.0149586396291852, 0.06188136339187622, 0.1311223804950714, 0.08191446959972382, 0.019586285576224327, -0.02480296604335308, -0.03388110175728798, -0.025523077696561813, -0.05937909707427025, 0.040128443390131, 0.06589099019765854, -0.16763372719287872, -0.039227183908224106, -0.09338314831256866, 0.09657008945941925, 0.0873042419552803, 0.06609832495450974, -0.1842060089111328, -0.008006223477423191, 0.08488986641168594, -0.03854905813932419, -0.13727426528930664, 0.09535189718008041, 0.01523482333868742, -0.15144726634025574, 0.03139317408204079, -0.04061909019947052, 0.12188644707202911, -0.07804752141237259, 0.09809603542089462, -0.08108244836330414, -0.07448557764291763, 0.02123199962079525, 0.1261177361011505, -0.30527687072753906, 0.20240111649036407, -0.0024993624538183212, -0.06486981362104416, -0.1243603527545929, -0.0032166161108762026, 0.002410882618278265, 0.07357452809810638, 0.10519039630889893, -0.007196315098553896, 0.001897757756523788, -0.06300821900367737, -0.01829923689365387, 0.032471053302288055, 0.13080233335494995, -0.0401318334043026, -0.021158374845981598, -0.050194524228572845, -0.001653497340157628, -0.03173094615340233, -0.06934895366430283, 0.02002747356891632, -0.19509181380271912, 0.08751901984214783, 0.04166261479258537, 0.09648149460554123, 0.029994789510965347, 0.004265148192644119, -0.09651939570903778, 0.24698667228221893, -0.07148019969463348, -0.10072879493236542, -0.10919588059186935, -0.046813901513814926, 0.03569883480668068, -0.05628936365246773, 0.04309194162487984, -0.0788632407784462, 0.028997479006648064, -0.06352769583463669, -0.19235502183437347, 0.12410202622413635, -0.09027006477117538, -0.04412810131907463, -0.02371402643620968, 0.2110891044139862, -0.05598580464720726, 0.010335659608244896, 0.02930437959730625, 0.01208863127976656, -0.11645778268575668, -0.09678568691015244, 0.031018631532788277, -0.007351789623498917, 0.050603240728378296, 0.041841957718133926, -0.05915454775094986, -0.017138581722974777, -0.052199993282556534, -0.022926922887563705, 0.3496883809566498, 0.14231905341148376, -0.043836336582899094, 0.19347235560417175, 0.12347975373268127, -0.07452994585037231, -0.3159443140029907, -0.1066238060593605, -0.10937739163637161, -0.04680149629712105, -0.07012093812227249, -0.2002030611038208, 0.06474938243627548, 0.00662544509395957, -0.013415241613984108, 0.12749312818050385, -0.2561831772327423, -0.07571036368608475, 0.15906259417533875, -0.017980827018618584, 0.3745945692062378, -0.1168576180934906, -0.10926306992769241, -0.03950892388820648, -0.14175476133823395, 0.16968177258968353, -0.01989765651524067, 0.11221715062856674, -0.009765521623194218, 0.14388824999332428, 0.05548359826207161, -0.023479344323277473, 0.08544106781482697, 0.004999885335564613, -0.03290518373250961, -0.10304180532693863, -0.05676887184381485, 0.007092386484146118, 0.02477436140179634, 0.018026655539870262, -0.041834570467472076, 0.02227151393890381, -0.11731979995965958, -0.04657655209302902, -0.08982590585947037, 0.04431166127324104, 0.03899754583835602, -0.07325074821710587, -0.002380647463724017, -0.07165111601352692, -0.012272949330508709, 0.022334342822432518, 0.20356793701648712, -0.08029330521821976, 0.16448934376239777, 0.09239562600851059, 0.12419285625219345, -0.14376309514045715, -0.00019283240544609725, -0.0762530043721199, -0.05611240118741989, 0.07737895101308823, -0.09433035552501678, 0.058893077075481415, 0.10901971161365509, -0.04567738622426987, 0.08828683942556381, 0.10377411544322968, 0.008936077356338501, 0.003213887568563223, 0.10916902124881744, -0.2667325437068939, -0.0296600554138422, -0.07532413303852081, 0.000883326749317348, 0.09092561900615692, 0.08562852442264557, 0.18840822577476501, 0.025361526757478714, -0.04293036088347435, -0.002770674182102084, 0.028597986325621605, -0.039021048694849014, 0.051667019724845886, 0.001123449532315135, 0.01947369985282421, -0.1530752182006836, 0.072522833943367, 0.01490565575659275, -0.15215420722961426, 0.021316176280379295, 0.16572684049606323, -0.11656328290700912, -0.1283872276544571, -0.06520111113786697, 0.08313824236392975, -0.11755692958831787, -0.01578943058848381, -0.03279297426342964, -0.13145680725574493, 0.07992171496152878, 0.12629036605358124, 0.05557859688997269, 0.0972496047616005, -0.06061713397502899, -0.020469192415475845, -0.018721895292401314, -0.014099318534135818, -0.012384648434817791, -0.007667020428925753, -0.055978111922740936, 0.0590752474963665, -0.026677248999476433, 0.1425808072090149, -0.09221141785383224, -0.1037059873342514, -0.16142144799232483, 0.0374140702188015, -0.11013076454401016, -0.08825794607400894, -0.08821134269237518, -0.050188567489385605, 0.002360827289521694, -0.019856395199894905, -0.04037635400891304, -0.05829505994915962, -0.12300454825162888, 0.0338277705013752, -0.040771447122097015, 0.024727050215005875, -0.07512269169092178, 0.015856385231018066, 0.08507686108350754, -0.03285100311040878, 0.15655414760112762, 0.1450488418340683, -0.1006515845656395, 0.10741901397705078, -0.14806775748729706, -0.09138492494821548, 0.11116421222686768, 0.015329592861235142, 0.0449691042304039, 0.09723787009716034, 0.013362943194806576, 0.0635865181684494, 0.032776717096567154, 0.05308786407113075, 0.027619892731308937, -0.11959987878799438, 0.06483134627342224, -0.03626115620136261, -0.14700546860694885, -0.049338050186634064, -0.05282869189977646, 0.01647452637553215, 0.013054544106125832, 0.09622690081596375, -0.05301849544048309, 0.10698331147432327, -0.04055701196193695, 0.0346808135509491, 0.017554637044668198, -0.1730053424835205, -0.03816922754049301, -0.08538098633289337, 0.03681723028421402, 0.014741539023816586, 0.25266793370246887, 0.030072299763560295, 0.012416383251547813, 0.032671261578798294, 0.08285367488861084, 0.03899408504366875, 0.010228337720036507, 0.17482228577136993, 0.1162426546216011, -0.06621865928173065, -0.10445023328065872, 0.0729617029428482, 0.016332454979419708, 0.01286179106682539, 0.13617953658103943, 0.008365051820874214, 0.005795429926365614, 0.08649782836437225, -0.016865963116288185, 0.009968153201043606, -0.10052056610584259, -0.13426925241947174, -0.022176474332809448, 0.05151832848787308, -0.04655967652797699, 0.11727844923734665, 0.1406494379043579, -0.01806013658642769, 0.03222079202532768, -0.021771740168333054, -0.05699979141354561, -0.1683429479598999, -0.1429590880870819, -0.06883849948644638, -0.13416796922683716, 0.00897989235818386, -0.11180389672517776, 0.05395037308335304, 0.06001098081469536, 0.06750501692295074, -0.06899319589138031, 0.10220931470394135, 0.04626858979463577, -0.11440542340278625, 0.06264589726924896, -0.0296088308095932, 0.09430401772260666, -0.02759445086121559, -0.019505485892295837, -0.09039592742919922, 0.014574515633285046, 0.011419114656746387, 0.06245238706469536, -0.04707273095846176, 0.007463190704584122, -0.14696238934993744, -0.08972041308879852, -0.0523175448179245, 0.0718572810292244, -0.050409089773893356, 0.14282815158367157, 0.00775480642914772, -0.0170906875282526, 0.039554283022880554, 0.22787313163280487, -0.07476283609867096, -0.04778539761900902, -0.05269690603017807, 0.20717895030975342, 0.02975541539490223, 0.1171872541308403, -0.022938819602131844, -0.006106364540755749, -0.0919521227478981, 0.3764844834804535, 0.30030161142349243, -0.09031439572572708, 0.011794124729931355, 0.02137952297925949, 0.04502861574292183, 0.1316293478012085, 0.1216534823179245, 0.10318691283464432, 0.3006802201271057, -0.07452366501092911, -0.04653361067175865, -0.012629742734134197, -0.023858042433857918, -0.09059546142816544, 0.1021224707365036, 0.04839762672781944, -0.06382183730602264, -0.03313443064689636, 0.0954432487487793, -0.25862133502960205, 0.1277991235256195, -0.12311873584985733, -0.17578600347042084, -0.06654827296733856, 0.009760108776390553, 0.10465722531080246, 0.015642458572983742, 0.0946015790104866, 0.007128213066607714, -0.11252258718013763, 0.06305865943431854, 0.03397420793771744, -0.22762253880500793, 0.0006893770187161863, 0.06642123311758041, -0.07006710022687912, -0.0024247700348496437, -0.026499588042497635, 0.05657242611050606, 0.0656052976846695, 0.054629553109407425, -0.00971333310008049, 0.03816632181406021, 0.0034184439573436975, -0.0585215799510479, 0.016623929142951965, 0.05121519789099693, 0.02472509816288948, -0.09763528406620026, 0.06927435845136642, -0.1574270874261856, 0.04766253009438515, -0.0030655991286039352, -0.04124255105853081, 0.006064958870410919, 0.008823691867291927, -0.06491616368293762, 0.05165379121899605, 0.07916834205389023, -0.0016257909592241049, -0.0062433634884655476, -0.057178743183612823, -0.02632102556526661, -0.027755750343203545, -0.09291748702526093, -0.10495562851428986, -0.14682936668395996, -0.11640441417694092, 0.09368976950645447, -0.01011267676949501, -0.1848134547472, 0.022154374048113823, -0.08606051653623581, 0.08319322764873505, -0.1670055389404297, 0.08040720224380493, 0.07041648775339127, 0.013038921169936657, -0.0031511052511632442, -0.02002427540719509, 0.054132770746946335, 0.086809903383255, -0.10407156497240067, -0.07400695979595184 ]
null
null
transformers
#DialoGPT medium model (Based on Harry Potter 1 through 4 plus 6, 18 epochs)
{"tags": ["conversational"]}
text-generation
Ayran/DialoGPT-medium-harry-potter-1-through-4-plus-6-e18
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
#DialoGPT medium model (Based on Harry Potter 1 through 4 plus 6, 18 epochs)
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.009697278961539268, 0.03208012506365776, -0.007204889785498381, 0.004809224978089333, 0.16726240515708923, 0.014898733235895634, 0.09765533357858658, 0.13672804832458496, -0.007841327227652073, -0.031050153076648712, 0.14490588009357452, 0.20411323010921478, -0.006439372431486845, 0.0661218985915184, -0.07572533935308456, -0.2683109939098358, 0.05759621039032936, 0.046649303287267685, 0.016515716910362244, 0.1200079694390297, 0.08573378622531891, -0.05473608896136284, 0.08714032918214798, -0.014583407901227474, -0.150366872549057, 0.017733458429574966, 0.043394338339567184, -0.12260226160287857, 0.11910516023635864, 0.05462685227394104, 0.07063519209623337, 0.014929565601050854, -0.07541623711585999, -0.1631229966878891, 0.03031250834465027, 0.01425902172923088, -0.0594632662832737, 0.04757995903491974, 0.059961482882499695, -0.10165371745824814, 0.10819483548402786, 0.09530027210712433, -0.013078106567263603, 0.06798283755779266, -0.16849711537361145, -0.020869607105851173, -0.01446688175201416, 0.009899779222905636, 0.05550243332982063, 0.09964893013238907, -0.03413357585668564, 0.10497362166643143, -0.09214533120393753, 0.11017382889986038, 0.10932035744190216, -0.32057443261146545, -0.005767723545432091, 0.09167823940515518, 0.039358653128147125, 0.07352814823389053, -0.04467793554067612, 0.06258884817361832, 0.018015462905168533, 0.017986174672842026, -0.014015024527907372, -0.07283061742782593, -0.11612214148044586, 0.04717336222529411, -0.08668071031570435, -0.059868961572647095, 0.2244078367948532, -0.05464440956711769, 0.06881742179393768, -0.05281897634267807, -0.10522868484258652, -0.04308144748210907, -0.029833965003490448, 0.00475557055324316, -0.07660607248544693, 0.08692064881324768, 0.00869679357856512, -0.09547875821590424, -0.1376667022705078, -0.02496783249080181, -0.1776352822780609, 0.16140350699424744, 0.02465328387916088, 0.05232657864689827, -0.2027255892753601, 0.09623090922832489, 0.017906051129102707, -0.08045592904090881, 0.022091427817940712, -0.10046248883008957, 0.029131146147847176, 0.013760408386588097, -0.04754498973488808, -0.061387211084365845, 0.0843690037727356, 0.11199145019054413, -0.01731434464454651, 0.025486016646027565, -0.039331406354904175, 0.08100687712430954, 0.03553595021367073, 0.09077847748994827, 0.007288969587534666, -0.028338588774204254, 0.025842782109975815, -0.13719046115875244, -0.003647835226729512, -0.07116208970546722, -0.16572439670562744, -0.021088803187012672, 0.02994808368384838, 0.08289173990488052, 0.015449047088623047, 0.11682453751564026, -0.03272046521306038, -0.025152435526251793, 0.03602350503206253, -0.047656361013650894, -0.012649794109165668, 0.016648368909955025, 0.013163427822291851, 0.12399329990148544, -0.0022096503525972366, 0.03235051408410072, -0.13653022050857544, 0.031423524022102356, -0.06793295592069626, -0.003740974934771657, -0.03486552834510803, -0.040637075901031494, 0.009043924510478973, -0.06862333416938782, 0.003486064961180091, -0.15030112862586975, -0.15063877403736115, 0.007587034720927477, -0.007836631499230862, -0.04107699543237686, -0.06370922178030014, -0.06952770054340363, -0.013550350442528725, 0.04251532256603241, -0.07093454152345657, -0.011352915316820145, -0.06403283774852753, 0.11004766076803207, -0.03197755664587021, 0.07921615242958069, -0.11953279376029968, 0.08390819281339645, -0.11260783672332764, -0.02386913076043129, -0.060801517218351364, 0.09317506104707718, -0.0006014376995153725, 0.09549830108880997, -0.006563255097717047, -0.017931854352355003, -0.07981178909540176, 0.06445012241601944, -0.042872510850429535, 0.21701598167419434, -0.0615808479487896, -0.11181682348251343, 0.28781595826148987, -0.052628401666879654, -0.1370542049407959, 0.11647392809391022, 0.008682746440172195, 0.05777018144726753, 0.10703510791063309, 0.19733482599258423, -0.015276194550096989, 0.004040541127324104, 0.09471915662288666, 0.11263324320316315, -0.11276852339506149, -0.033160366117954254, 0.013019153848290443, -0.04081077128648758, -0.10867965966463089, 0.04689536616206169, 0.09810488671064377, 0.07090286910533905, -0.04786505550146103, -0.03377414867281914, -0.01366397924721241, 0.0052589005790650845, 0.08885077387094498, -0.007157256826758385, 0.10962837189435959, -0.05819983780384064, -0.03796621412038803, -0.029282379895448685, -0.012126247398555279, -0.03951939567923546, 0.03137664496898651, -0.043376367539167404, 0.10821941494941711, -0.011204327456653118, 0.06364280730485916, -0.16185984015464783, -0.07691477984189987, -0.017002692446112633, 0.1581239402294159, 0.024538565427064896, 0.09859629720449448, 0.0552486926317215, -0.040398042649030685, -0.0012767292791977525, 0.012792680412530899, 0.15581141412258148, -0.022091681137681007, -0.065607450902462, -0.052166227251291275, 0.08642971515655518, -0.05641226842999458, 0.04504093527793884, -0.05937713757157326, 0.012367865070700645, 0.05064384639263153, 0.10342344641685486, -0.00018274025933351368, 0.03323284164071083, -0.008164864964783192, 0.002145637758076191, -0.058205123990774155, 0.007405933458358049, 0.10799351334571838, 0.00036868182360194623, -0.07365862280130386, 0.22074243426322937, -0.17796069383621216, 0.1765957772731781, 0.1893044263124466, -0.299345999956131, 0.017949223518371582, -0.10759581625461578, -0.04561871662735939, 0.014407722279429436, 0.05567655712366104, -0.0454222597181797, 0.1703362911939621, -0.009871348738670349, 0.18874616920948029, -0.04946064203977585, -0.04464937001466751, -0.0200483538210392, -0.05118836089968681, -0.0024189651012420654, 0.07781197130680084, 0.10685696452856064, -0.13992026448249817, 0.1964332014322281, 0.1621224284172058, 0.048237916082143784, 0.19945049285888672, 0.015346456319093704, -0.011589210480451584, 0.0909530371427536, 0.005220826715230942, -0.058739423751831055, -0.07409929484128952, -0.2594851851463318, -0.030033592134714127, 0.07992640137672424, 0.0422382652759552, 0.1212305948138237, -0.11349532753229141, -0.038956157863140106, -0.01763172075152397, -0.023146281018853188, 0.021672505885362625, 0.0914369598031044, 0.06075398623943329, 0.13201528787612915, -0.001710098935291171, -0.007300339173525572, 0.10524573177099228, 0.01783694699406624, -0.09354141354560852, 0.18308524787425995, -0.13652534782886505, -0.37097251415252686, -0.13911493122577667, -0.18057456612586975, -0.05449081212282181, 0.05712554603815079, 0.11679314076900482, -0.12011238187551498, -0.018752124160528183, 0.01578843593597412, 0.10931742936372757, -0.08449502289295197, 0.0021454424131661654, -0.06880278885364532, 0.0321490578353405, -0.10310184955596924, -0.09194442629814148, -0.055416494607925415, -0.031392451375722885, -0.08001253753900528, 0.1423761546611786, -0.10777941346168518, 0.04476889222860336, 0.20262959599494934, 0.04653622955083847, 0.05625178664922714, -0.044105201959609985, 0.19377262890338898, -0.11264272034168243, -0.01661740615963936, 0.19215328991413116, -0.048360925167798996, 0.07476246356964111, 0.1232115849852562, -0.006348740309476852, -0.08765771239995956, 0.03011748194694519, -0.02085109055042267, -0.07988511025905609, -0.23219464719295502, -0.13938382267951965, -0.12429051846265793, 0.09477275609970093, 0.028005298227071762, 0.056365787982940674, 0.17219258844852448, 0.06577219814062119, -0.038416244089603424, 0.006410336587578058, 0.02959546446800232, 0.08237514644861221, 0.23417828977108002, -0.06035616248846054, 0.1364797055721283, -0.03420931473374367, -0.14982740581035614, 0.08169995993375778, 0.0713929831981659, 0.10213395953178406, 0.06678459793329239, 0.0804823637008667, 0.0149586396291852, 0.06188136339187622, 0.1311223804950714, 0.08191446959972382, 0.019586285576224327, -0.02480296604335308, -0.03388110175728798, -0.025523077696561813, -0.05937909707427025, 0.040128443390131, 0.06589099019765854, -0.16763372719287872, -0.039227183908224106, -0.09338314831256866, 0.09657008945941925, 0.0873042419552803, 0.06609832495450974, -0.1842060089111328, -0.008006223477423191, 0.08488986641168594, -0.03854905813932419, -0.13727426528930664, 0.09535189718008041, 0.01523482333868742, -0.15144726634025574, 0.03139317408204079, -0.04061909019947052, 0.12188644707202911, -0.07804752141237259, 0.09809603542089462, -0.08108244836330414, -0.07448557764291763, 0.02123199962079525, 0.1261177361011505, -0.30527687072753906, 0.20240111649036407, -0.0024993624538183212, -0.06486981362104416, -0.1243603527545929, -0.0032166161108762026, 0.002410882618278265, 0.07357452809810638, 0.10519039630889893, -0.007196315098553896, 0.001897757756523788, -0.06300821900367737, -0.01829923689365387, 0.032471053302288055, 0.13080233335494995, -0.0401318334043026, -0.021158374845981598, -0.050194524228572845, -0.001653497340157628, -0.03173094615340233, -0.06934895366430283, 0.02002747356891632, -0.19509181380271912, 0.08751901984214783, 0.04166261479258537, 0.09648149460554123, 0.029994789510965347, 0.004265148192644119, -0.09651939570903778, 0.24698667228221893, -0.07148019969463348, -0.10072879493236542, -0.10919588059186935, -0.046813901513814926, 0.03569883480668068, -0.05628936365246773, 0.04309194162487984, -0.0788632407784462, 0.028997479006648064, -0.06352769583463669, -0.19235502183437347, 0.12410202622413635, -0.09027006477117538, -0.04412810131907463, -0.02371402643620968, 0.2110891044139862, -0.05598580464720726, 0.010335659608244896, 0.02930437959730625, 0.01208863127976656, -0.11645778268575668, -0.09678568691015244, 0.031018631532788277, -0.007351789623498917, 0.050603240728378296, 0.041841957718133926, -0.05915454775094986, -0.017138581722974777, -0.052199993282556534, -0.022926922887563705, 0.3496883809566498, 0.14231905341148376, -0.043836336582899094, 0.19347235560417175, 0.12347975373268127, -0.07452994585037231, -0.3159443140029907, -0.1066238060593605, -0.10937739163637161, -0.04680149629712105, -0.07012093812227249, -0.2002030611038208, 0.06474938243627548, 0.00662544509395957, -0.013415241613984108, 0.12749312818050385, -0.2561831772327423, -0.07571036368608475, 0.15906259417533875, -0.017980827018618584, 0.3745945692062378, -0.1168576180934906, -0.10926306992769241, -0.03950892388820648, -0.14175476133823395, 0.16968177258968353, -0.01989765651524067, 0.11221715062856674, -0.009765521623194218, 0.14388824999332428, 0.05548359826207161, -0.023479344323277473, 0.08544106781482697, 0.004999885335564613, -0.03290518373250961, -0.10304180532693863, -0.05676887184381485, 0.007092386484146118, 0.02477436140179634, 0.018026655539870262, -0.041834570467472076, 0.02227151393890381, -0.11731979995965958, -0.04657655209302902, -0.08982590585947037, 0.04431166127324104, 0.03899754583835602, -0.07325074821710587, -0.002380647463724017, -0.07165111601352692, -0.012272949330508709, 0.022334342822432518, 0.20356793701648712, -0.08029330521821976, 0.16448934376239777, 0.09239562600851059, 0.12419285625219345, -0.14376309514045715, -0.00019283240544609725, -0.0762530043721199, -0.05611240118741989, 0.07737895101308823, -0.09433035552501678, 0.058893077075481415, 0.10901971161365509, -0.04567738622426987, 0.08828683942556381, 0.10377411544322968, 0.008936077356338501, 0.003213887568563223, 0.10916902124881744, -0.2667325437068939, -0.0296600554138422, -0.07532413303852081, 0.000883326749317348, 0.09092561900615692, 0.08562852442264557, 0.18840822577476501, 0.025361526757478714, -0.04293036088347435, -0.002770674182102084, 0.028597986325621605, -0.039021048694849014, 0.051667019724845886, 0.001123449532315135, 0.01947369985282421, -0.1530752182006836, 0.072522833943367, 0.01490565575659275, -0.15215420722961426, 0.021316176280379295, 0.16572684049606323, -0.11656328290700912, -0.1283872276544571, -0.06520111113786697, 0.08313824236392975, -0.11755692958831787, -0.01578943058848381, -0.03279297426342964, -0.13145680725574493, 0.07992171496152878, 0.12629036605358124, 0.05557859688997269, 0.0972496047616005, -0.06061713397502899, -0.020469192415475845, -0.018721895292401314, -0.014099318534135818, -0.012384648434817791, -0.007667020428925753, -0.055978111922740936, 0.0590752474963665, -0.026677248999476433, 0.1425808072090149, -0.09221141785383224, -0.1037059873342514, -0.16142144799232483, 0.0374140702188015, -0.11013076454401016, -0.08825794607400894, -0.08821134269237518, -0.050188567489385605, 0.002360827289521694, -0.019856395199894905, -0.04037635400891304, -0.05829505994915962, -0.12300454825162888, 0.0338277705013752, -0.040771447122097015, 0.024727050215005875, -0.07512269169092178, 0.015856385231018066, 0.08507686108350754, -0.03285100311040878, 0.15655414760112762, 0.1450488418340683, -0.1006515845656395, 0.10741901397705078, -0.14806775748729706, -0.09138492494821548, 0.11116421222686768, 0.015329592861235142, 0.0449691042304039, 0.09723787009716034, 0.013362943194806576, 0.0635865181684494, 0.032776717096567154, 0.05308786407113075, 0.027619892731308937, -0.11959987878799438, 0.06483134627342224, -0.03626115620136261, -0.14700546860694885, -0.049338050186634064, -0.05282869189977646, 0.01647452637553215, 0.013054544106125832, 0.09622690081596375, -0.05301849544048309, 0.10698331147432327, -0.04055701196193695, 0.0346808135509491, 0.017554637044668198, -0.1730053424835205, -0.03816922754049301, -0.08538098633289337, 0.03681723028421402, 0.014741539023816586, 0.25266793370246887, 0.030072299763560295, 0.012416383251547813, 0.032671261578798294, 0.08285367488861084, 0.03899408504366875, 0.010228337720036507, 0.17482228577136993, 0.1162426546216011, -0.06621865928173065, -0.10445023328065872, 0.0729617029428482, 0.016332454979419708, 0.01286179106682539, 0.13617953658103943, 0.008365051820874214, 0.005795429926365614, 0.08649782836437225, -0.016865963116288185, 0.009968153201043606, -0.10052056610584259, -0.13426925241947174, -0.022176474332809448, 0.05151832848787308, -0.04655967652797699, 0.11727844923734665, 0.1406494379043579, -0.01806013658642769, 0.03222079202532768, -0.021771740168333054, -0.05699979141354561, -0.1683429479598999, -0.1429590880870819, -0.06883849948644638, -0.13416796922683716, 0.00897989235818386, -0.11180389672517776, 0.05395037308335304, 0.06001098081469536, 0.06750501692295074, -0.06899319589138031, 0.10220931470394135, 0.04626858979463577, -0.11440542340278625, 0.06264589726924896, -0.0296088308095932, 0.09430401772260666, -0.02759445086121559, -0.019505485892295837, -0.09039592742919922, 0.014574515633285046, 0.011419114656746387, 0.06245238706469536, -0.04707273095846176, 0.007463190704584122, -0.14696238934993744, -0.08972041308879852, -0.0523175448179245, 0.0718572810292244, -0.050409089773893356, 0.14282815158367157, 0.00775480642914772, -0.0170906875282526, 0.039554283022880554, 0.22787313163280487, -0.07476283609867096, -0.04778539761900902, -0.05269690603017807, 0.20717895030975342, 0.02975541539490223, 0.1171872541308403, -0.022938819602131844, -0.006106364540755749, -0.0919521227478981, 0.3764844834804535, 0.30030161142349243, -0.09031439572572708, 0.011794124729931355, 0.02137952297925949, 0.04502861574292183, 0.1316293478012085, 0.1216534823179245, 0.10318691283464432, 0.3006802201271057, -0.07452366501092911, -0.04653361067175865, -0.012629742734134197, -0.023858042433857918, -0.09059546142816544, 0.1021224707365036, 0.04839762672781944, -0.06382183730602264, -0.03313443064689636, 0.0954432487487793, -0.25862133502960205, 0.1277991235256195, -0.12311873584985733, -0.17578600347042084, -0.06654827296733856, 0.009760108776390553, 0.10465722531080246, 0.015642458572983742, 0.0946015790104866, 0.007128213066607714, -0.11252258718013763, 0.06305865943431854, 0.03397420793771744, -0.22762253880500793, 0.0006893770187161863, 0.06642123311758041, -0.07006710022687912, -0.0024247700348496437, -0.026499588042497635, 0.05657242611050606, 0.0656052976846695, 0.054629553109407425, -0.00971333310008049, 0.03816632181406021, 0.0034184439573436975, -0.0585215799510479, 0.016623929142951965, 0.05121519789099693, 0.02472509816288948, -0.09763528406620026, 0.06927435845136642, -0.1574270874261856, 0.04766253009438515, -0.0030655991286039352, -0.04124255105853081, 0.006064958870410919, 0.008823691867291927, -0.06491616368293762, 0.05165379121899605, 0.07916834205389023, -0.0016257909592241049, -0.0062433634884655476, -0.057178743183612823, -0.02632102556526661, -0.027755750343203545, -0.09291748702526093, -0.10495562851428986, -0.14682936668395996, -0.11640441417694092, 0.09368976950645447, -0.01011267676949501, -0.1848134547472, 0.022154374048113823, -0.08606051653623581, 0.08319322764873505, -0.1670055389404297, 0.08040720224380493, 0.07041648775339127, 0.013038921169936657, -0.0031511052511632442, -0.02002427540719509, 0.054132770746946335, 0.086809903383255, -0.10407156497240067, -0.07400695979595184 ]
null
null
transformers
#DialoGPT medium model (Harry Potter 1 through 4 plus 6)
{"tags": ["conversational"]}
text-generation
Ayran/DialoGPT-medium-harry-potter-1-through-4-plus-6
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
#DialoGPT medium model (Harry Potter 1 through 4 plus 6)
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.009697278961539268, 0.03208012506365776, -0.007204889785498381, 0.004809224978089333, 0.16726240515708923, 0.014898733235895634, 0.09765533357858658, 0.13672804832458496, -0.007841327227652073, -0.031050153076648712, 0.14490588009357452, 0.20411323010921478, -0.006439372431486845, 0.0661218985915184, -0.07572533935308456, -0.2683109939098358, 0.05759621039032936, 0.046649303287267685, 0.016515716910362244, 0.1200079694390297, 0.08573378622531891, -0.05473608896136284, 0.08714032918214798, -0.014583407901227474, -0.150366872549057, 0.017733458429574966, 0.043394338339567184, -0.12260226160287857, 0.11910516023635864, 0.05462685227394104, 0.07063519209623337, 0.014929565601050854, -0.07541623711585999, -0.1631229966878891, 0.03031250834465027, 0.01425902172923088, -0.0594632662832737, 0.04757995903491974, 0.059961482882499695, -0.10165371745824814, 0.10819483548402786, 0.09530027210712433, -0.013078106567263603, 0.06798283755779266, -0.16849711537361145, -0.020869607105851173, -0.01446688175201416, 0.009899779222905636, 0.05550243332982063, 0.09964893013238907, -0.03413357585668564, 0.10497362166643143, -0.09214533120393753, 0.11017382889986038, 0.10932035744190216, -0.32057443261146545, -0.005767723545432091, 0.09167823940515518, 0.039358653128147125, 0.07352814823389053, -0.04467793554067612, 0.06258884817361832, 0.018015462905168533, 0.017986174672842026, -0.014015024527907372, -0.07283061742782593, -0.11612214148044586, 0.04717336222529411, -0.08668071031570435, -0.059868961572647095, 0.2244078367948532, -0.05464440956711769, 0.06881742179393768, -0.05281897634267807, -0.10522868484258652, -0.04308144748210907, -0.029833965003490448, 0.00475557055324316, -0.07660607248544693, 0.08692064881324768, 0.00869679357856512, -0.09547875821590424, -0.1376667022705078, -0.02496783249080181, -0.1776352822780609, 0.16140350699424744, 0.02465328387916088, 0.05232657864689827, -0.2027255892753601, 0.09623090922832489, 0.017906051129102707, -0.08045592904090881, 0.022091427817940712, -0.10046248883008957, 0.029131146147847176, 0.013760408386588097, -0.04754498973488808, -0.061387211084365845, 0.0843690037727356, 0.11199145019054413, -0.01731434464454651, 0.025486016646027565, -0.039331406354904175, 0.08100687712430954, 0.03553595021367073, 0.09077847748994827, 0.007288969587534666, -0.028338588774204254, 0.025842782109975815, -0.13719046115875244, -0.003647835226729512, -0.07116208970546722, -0.16572439670562744, -0.021088803187012672, 0.02994808368384838, 0.08289173990488052, 0.015449047088623047, 0.11682453751564026, -0.03272046521306038, -0.025152435526251793, 0.03602350503206253, -0.047656361013650894, -0.012649794109165668, 0.016648368909955025, 0.013163427822291851, 0.12399329990148544, -0.0022096503525972366, 0.03235051408410072, -0.13653022050857544, 0.031423524022102356, -0.06793295592069626, -0.003740974934771657, -0.03486552834510803, -0.040637075901031494, 0.009043924510478973, -0.06862333416938782, 0.003486064961180091, -0.15030112862586975, -0.15063877403736115, 0.007587034720927477, -0.007836631499230862, -0.04107699543237686, -0.06370922178030014, -0.06952770054340363, -0.013550350442528725, 0.04251532256603241, -0.07093454152345657, -0.011352915316820145, -0.06403283774852753, 0.11004766076803207, -0.03197755664587021, 0.07921615242958069, -0.11953279376029968, 0.08390819281339645, -0.11260783672332764, -0.02386913076043129, -0.060801517218351364, 0.09317506104707718, -0.0006014376995153725, 0.09549830108880997, -0.006563255097717047, -0.017931854352355003, -0.07981178909540176, 0.06445012241601944, -0.042872510850429535, 0.21701598167419434, -0.0615808479487896, -0.11181682348251343, 0.28781595826148987, -0.052628401666879654, -0.1370542049407959, 0.11647392809391022, 0.008682746440172195, 0.05777018144726753, 0.10703510791063309, 0.19733482599258423, -0.015276194550096989, 0.004040541127324104, 0.09471915662288666, 0.11263324320316315, -0.11276852339506149, -0.033160366117954254, 0.013019153848290443, -0.04081077128648758, -0.10867965966463089, 0.04689536616206169, 0.09810488671064377, 0.07090286910533905, -0.04786505550146103, -0.03377414867281914, -0.01366397924721241, 0.0052589005790650845, 0.08885077387094498, -0.007157256826758385, 0.10962837189435959, -0.05819983780384064, -0.03796621412038803, -0.029282379895448685, -0.012126247398555279, -0.03951939567923546, 0.03137664496898651, -0.043376367539167404, 0.10821941494941711, -0.011204327456653118, 0.06364280730485916, -0.16185984015464783, -0.07691477984189987, -0.017002692446112633, 0.1581239402294159, 0.024538565427064896, 0.09859629720449448, 0.0552486926317215, -0.040398042649030685, -0.0012767292791977525, 0.012792680412530899, 0.15581141412258148, -0.022091681137681007, -0.065607450902462, -0.052166227251291275, 0.08642971515655518, -0.05641226842999458, 0.04504093527793884, -0.05937713757157326, 0.012367865070700645, 0.05064384639263153, 0.10342344641685486, -0.00018274025933351368, 0.03323284164071083, -0.008164864964783192, 0.002145637758076191, -0.058205123990774155, 0.007405933458358049, 0.10799351334571838, 0.00036868182360194623, -0.07365862280130386, 0.22074243426322937, -0.17796069383621216, 0.1765957772731781, 0.1893044263124466, -0.299345999956131, 0.017949223518371582, -0.10759581625461578, -0.04561871662735939, 0.014407722279429436, 0.05567655712366104, -0.0454222597181797, 0.1703362911939621, -0.009871348738670349, 0.18874616920948029, -0.04946064203977585, -0.04464937001466751, -0.0200483538210392, -0.05118836089968681, -0.0024189651012420654, 0.07781197130680084, 0.10685696452856064, -0.13992026448249817, 0.1964332014322281, 0.1621224284172058, 0.048237916082143784, 0.19945049285888672, 0.015346456319093704, -0.011589210480451584, 0.0909530371427536, 0.005220826715230942, -0.058739423751831055, -0.07409929484128952, -0.2594851851463318, -0.030033592134714127, 0.07992640137672424, 0.0422382652759552, 0.1212305948138237, -0.11349532753229141, -0.038956157863140106, -0.01763172075152397, -0.023146281018853188, 0.021672505885362625, 0.0914369598031044, 0.06075398623943329, 0.13201528787612915, -0.001710098935291171, -0.007300339173525572, 0.10524573177099228, 0.01783694699406624, -0.09354141354560852, 0.18308524787425995, -0.13652534782886505, -0.37097251415252686, -0.13911493122577667, -0.18057456612586975, -0.05449081212282181, 0.05712554603815079, 0.11679314076900482, -0.12011238187551498, -0.018752124160528183, 0.01578843593597412, 0.10931742936372757, -0.08449502289295197, 0.0021454424131661654, -0.06880278885364532, 0.0321490578353405, -0.10310184955596924, -0.09194442629814148, -0.055416494607925415, -0.031392451375722885, -0.08001253753900528, 0.1423761546611786, -0.10777941346168518, 0.04476889222860336, 0.20262959599494934, 0.04653622955083847, 0.05625178664922714, -0.044105201959609985, 0.19377262890338898, -0.11264272034168243, -0.01661740615963936, 0.19215328991413116, -0.048360925167798996, 0.07476246356964111, 0.1232115849852562, -0.006348740309476852, -0.08765771239995956, 0.03011748194694519, -0.02085109055042267, -0.07988511025905609, -0.23219464719295502, -0.13938382267951965, -0.12429051846265793, 0.09477275609970093, 0.028005298227071762, 0.056365787982940674, 0.17219258844852448, 0.06577219814062119, -0.038416244089603424, 0.006410336587578058, 0.02959546446800232, 0.08237514644861221, 0.23417828977108002, -0.06035616248846054, 0.1364797055721283, -0.03420931473374367, -0.14982740581035614, 0.08169995993375778, 0.0713929831981659, 0.10213395953178406, 0.06678459793329239, 0.0804823637008667, 0.0149586396291852, 0.06188136339187622, 0.1311223804950714, 0.08191446959972382, 0.019586285576224327, -0.02480296604335308, -0.03388110175728798, -0.025523077696561813, -0.05937909707427025, 0.040128443390131, 0.06589099019765854, -0.16763372719287872, -0.039227183908224106, -0.09338314831256866, 0.09657008945941925, 0.0873042419552803, 0.06609832495450974, -0.1842060089111328, -0.008006223477423191, 0.08488986641168594, -0.03854905813932419, -0.13727426528930664, 0.09535189718008041, 0.01523482333868742, -0.15144726634025574, 0.03139317408204079, -0.04061909019947052, 0.12188644707202911, -0.07804752141237259, 0.09809603542089462, -0.08108244836330414, -0.07448557764291763, 0.02123199962079525, 0.1261177361011505, -0.30527687072753906, 0.20240111649036407, -0.0024993624538183212, -0.06486981362104416, -0.1243603527545929, -0.0032166161108762026, 0.002410882618278265, 0.07357452809810638, 0.10519039630889893, -0.007196315098553896, 0.001897757756523788, -0.06300821900367737, -0.01829923689365387, 0.032471053302288055, 0.13080233335494995, -0.0401318334043026, -0.021158374845981598, -0.050194524228572845, -0.001653497340157628, -0.03173094615340233, -0.06934895366430283, 0.02002747356891632, -0.19509181380271912, 0.08751901984214783, 0.04166261479258537, 0.09648149460554123, 0.029994789510965347, 0.004265148192644119, -0.09651939570903778, 0.24698667228221893, -0.07148019969463348, -0.10072879493236542, -0.10919588059186935, -0.046813901513814926, 0.03569883480668068, -0.05628936365246773, 0.04309194162487984, -0.0788632407784462, 0.028997479006648064, -0.06352769583463669, -0.19235502183437347, 0.12410202622413635, -0.09027006477117538, -0.04412810131907463, -0.02371402643620968, 0.2110891044139862, -0.05598580464720726, 0.010335659608244896, 0.02930437959730625, 0.01208863127976656, -0.11645778268575668, -0.09678568691015244, 0.031018631532788277, -0.007351789623498917, 0.050603240728378296, 0.041841957718133926, -0.05915454775094986, -0.017138581722974777, -0.052199993282556534, -0.022926922887563705, 0.3496883809566498, 0.14231905341148376, -0.043836336582899094, 0.19347235560417175, 0.12347975373268127, -0.07452994585037231, -0.3159443140029907, -0.1066238060593605, -0.10937739163637161, -0.04680149629712105, -0.07012093812227249, -0.2002030611038208, 0.06474938243627548, 0.00662544509395957, -0.013415241613984108, 0.12749312818050385, -0.2561831772327423, -0.07571036368608475, 0.15906259417533875, -0.017980827018618584, 0.3745945692062378, -0.1168576180934906, -0.10926306992769241, -0.03950892388820648, -0.14175476133823395, 0.16968177258968353, -0.01989765651524067, 0.11221715062856674, -0.009765521623194218, 0.14388824999332428, 0.05548359826207161, -0.023479344323277473, 0.08544106781482697, 0.004999885335564613, -0.03290518373250961, -0.10304180532693863, -0.05676887184381485, 0.007092386484146118, 0.02477436140179634, 0.018026655539870262, -0.041834570467472076, 0.02227151393890381, -0.11731979995965958, -0.04657655209302902, -0.08982590585947037, 0.04431166127324104, 0.03899754583835602, -0.07325074821710587, -0.002380647463724017, -0.07165111601352692, -0.012272949330508709, 0.022334342822432518, 0.20356793701648712, -0.08029330521821976, 0.16448934376239777, 0.09239562600851059, 0.12419285625219345, -0.14376309514045715, -0.00019283240544609725, -0.0762530043721199, -0.05611240118741989, 0.07737895101308823, -0.09433035552501678, 0.058893077075481415, 0.10901971161365509, -0.04567738622426987, 0.08828683942556381, 0.10377411544322968, 0.008936077356338501, 0.003213887568563223, 0.10916902124881744, -0.2667325437068939, -0.0296600554138422, -0.07532413303852081, 0.000883326749317348, 0.09092561900615692, 0.08562852442264557, 0.18840822577476501, 0.025361526757478714, -0.04293036088347435, -0.002770674182102084, 0.028597986325621605, -0.039021048694849014, 0.051667019724845886, 0.001123449532315135, 0.01947369985282421, -0.1530752182006836, 0.072522833943367, 0.01490565575659275, -0.15215420722961426, 0.021316176280379295, 0.16572684049606323, -0.11656328290700912, -0.1283872276544571, -0.06520111113786697, 0.08313824236392975, -0.11755692958831787, -0.01578943058848381, -0.03279297426342964, -0.13145680725574493, 0.07992171496152878, 0.12629036605358124, 0.05557859688997269, 0.0972496047616005, -0.06061713397502899, -0.020469192415475845, -0.018721895292401314, -0.014099318534135818, -0.012384648434817791, -0.007667020428925753, -0.055978111922740936, 0.0590752474963665, -0.026677248999476433, 0.1425808072090149, -0.09221141785383224, -0.1037059873342514, -0.16142144799232483, 0.0374140702188015, -0.11013076454401016, -0.08825794607400894, -0.08821134269237518, -0.050188567489385605, 0.002360827289521694, -0.019856395199894905, -0.04037635400891304, -0.05829505994915962, -0.12300454825162888, 0.0338277705013752, -0.040771447122097015, 0.024727050215005875, -0.07512269169092178, 0.015856385231018066, 0.08507686108350754, -0.03285100311040878, 0.15655414760112762, 0.1450488418340683, -0.1006515845656395, 0.10741901397705078, -0.14806775748729706, -0.09138492494821548, 0.11116421222686768, 0.015329592861235142, 0.0449691042304039, 0.09723787009716034, 0.013362943194806576, 0.0635865181684494, 0.032776717096567154, 0.05308786407113075, 0.027619892731308937, -0.11959987878799438, 0.06483134627342224, -0.03626115620136261, -0.14700546860694885, -0.049338050186634064, -0.05282869189977646, 0.01647452637553215, 0.013054544106125832, 0.09622690081596375, -0.05301849544048309, 0.10698331147432327, -0.04055701196193695, 0.0346808135509491, 0.017554637044668198, -0.1730053424835205, -0.03816922754049301, -0.08538098633289337, 0.03681723028421402, 0.014741539023816586, 0.25266793370246887, 0.030072299763560295, 0.012416383251547813, 0.032671261578798294, 0.08285367488861084, 0.03899408504366875, 0.010228337720036507, 0.17482228577136993, 0.1162426546216011, -0.06621865928173065, -0.10445023328065872, 0.0729617029428482, 0.016332454979419708, 0.01286179106682539, 0.13617953658103943, 0.008365051820874214, 0.005795429926365614, 0.08649782836437225, -0.016865963116288185, 0.009968153201043606, -0.10052056610584259, -0.13426925241947174, -0.022176474332809448, 0.05151832848787308, -0.04655967652797699, 0.11727844923734665, 0.1406494379043579, -0.01806013658642769, 0.03222079202532768, -0.021771740168333054, -0.05699979141354561, -0.1683429479598999, -0.1429590880870819, -0.06883849948644638, -0.13416796922683716, 0.00897989235818386, -0.11180389672517776, 0.05395037308335304, 0.06001098081469536, 0.06750501692295074, -0.06899319589138031, 0.10220931470394135, 0.04626858979463577, -0.11440542340278625, 0.06264589726924896, -0.0296088308095932, 0.09430401772260666, -0.02759445086121559, -0.019505485892295837, -0.09039592742919922, 0.014574515633285046, 0.011419114656746387, 0.06245238706469536, -0.04707273095846176, 0.007463190704584122, -0.14696238934993744, -0.08972041308879852, -0.0523175448179245, 0.0718572810292244, -0.050409089773893356, 0.14282815158367157, 0.00775480642914772, -0.0170906875282526, 0.039554283022880554, 0.22787313163280487, -0.07476283609867096, -0.04778539761900902, -0.05269690603017807, 0.20717895030975342, 0.02975541539490223, 0.1171872541308403, -0.022938819602131844, -0.006106364540755749, -0.0919521227478981, 0.3764844834804535, 0.30030161142349243, -0.09031439572572708, 0.011794124729931355, 0.02137952297925949, 0.04502861574292183, 0.1316293478012085, 0.1216534823179245, 0.10318691283464432, 0.3006802201271057, -0.07452366501092911, -0.04653361067175865, -0.012629742734134197, -0.023858042433857918, -0.09059546142816544, 0.1021224707365036, 0.04839762672781944, -0.06382183730602264, -0.03313443064689636, 0.0954432487487793, -0.25862133502960205, 0.1277991235256195, -0.12311873584985733, -0.17578600347042084, -0.06654827296733856, 0.009760108776390553, 0.10465722531080246, 0.015642458572983742, 0.0946015790104866, 0.007128213066607714, -0.11252258718013763, 0.06305865943431854, 0.03397420793771744, -0.22762253880500793, 0.0006893770187161863, 0.06642123311758041, -0.07006710022687912, -0.0024247700348496437, -0.026499588042497635, 0.05657242611050606, 0.0656052976846695, 0.054629553109407425, -0.00971333310008049, 0.03816632181406021, 0.0034184439573436975, -0.0585215799510479, 0.016623929142951965, 0.05121519789099693, 0.02472509816288948, -0.09763528406620026, 0.06927435845136642, -0.1574270874261856, 0.04766253009438515, -0.0030655991286039352, -0.04124255105853081, 0.006064958870410919, 0.008823691867291927, -0.06491616368293762, 0.05165379121899605, 0.07916834205389023, -0.0016257909592241049, -0.0062433634884655476, -0.057178743183612823, -0.02632102556526661, -0.027755750343203545, -0.09291748702526093, -0.10495562851428986, -0.14682936668395996, -0.11640441417694092, 0.09368976950645447, -0.01011267676949501, -0.1848134547472, 0.022154374048113823, -0.08606051653623581, 0.08319322764873505, -0.1670055389404297, 0.08040720224380493, 0.07041648775339127, 0.013038921169936657, -0.0031511052511632442, -0.02002427540719509, 0.054132770746946335, 0.086809903383255, -0.10407156497240067, -0.07400695979595184 ]
null
null
transformers
# Gandalf DialoGPT Model
{"tags": ["conversational"]}
text-generation
Ayran/DialoGPT-small-gandalf
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Gandalf DialoGPT Model
[ "# Gandalf DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Gandalf DialoGPT Model" ]
[ 51, 9 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Gandalf DialoGPT Model" ]
[ -0.00460621016100049, 0.1131497472524643, -0.007377896923571825, 0.10403112322092056, 0.1234591081738472, 0.0053472090512514114, 0.16734673082828522, 0.11113400012254715, -0.024191226810216904, -0.050235453993082047, 0.14147472381591797, 0.17374832928180695, 0.011733937077224255, 0.022198386490345, -0.029490143060684204, -0.3504922091960907, 0.035186078399419785, 0.026421932503581047, -0.0921204686164856, 0.10529574751853943, 0.07343821972608566, -0.007619699463248253, 0.08505943417549133, -0.00766177661716938, -0.10638485103845596, -0.004496118985116482, -0.009203296154737473, -0.09580175578594208, 0.1542641818523407, 0.05589836835861206, -0.011733349412679672, -0.023313459008932114, -0.06592471897602081, -0.1270531266927719, 0.04724227637052536, -0.030675917863845825, -0.017084665596485138, 0.03495916724205017, 0.015342853963375092, -0.043434564024209976, 0.18884336948394775, 0.13564030826091766, 0.004588625393807888, 0.025735430419445038, -0.13440880179405212, -0.06755173206329346, 0.02742592804133892, 0.0679701566696167, -0.0006592888385057449, 0.09048527479171753, -0.02265041507780552, 0.11403529345989227, -0.016243746504187584, 0.10664872825145721, 0.11125865578651428, -0.39968055486679077, -0.05148051306605339, 0.17660769820213318, 0.04157522693276405, 0.12396489083766937, -0.081522636115551, 0.07938007265329361, -0.013491660356521606, 0.0323554091155529, -0.03741776943206787, -0.09615781158208847, -0.11795090138912201, 0.04023265838623047, -0.1091466099023819, 0.035852015018463135, 0.19416549801826477, -0.0997980535030365, 0.05032339319586754, -0.060295119881629944, -0.0821363776922226, 0.014219047501683235, -0.022646615281701088, -0.06886877864599228, -0.06267353892326355, 0.07074429839849472, 0.02563374862074852, -0.10338679701089859, -0.09132552146911621, -0.015044694766402245, -0.13389140367507935, 0.12590542435646057, 0.05910893529653549, 0.0581275075674057, -0.21681106090545654, 0.11824807524681091, -0.03351021930575371, -0.06960809230804443, -0.021827291697263718, -0.07281529158353806, -0.009254757314920425, 0.00857630092650652, 0.007253926247358322, -0.011432142928242683, 0.03155769035220146, 0.19284285604953766, 0.012689546681940556, 0.033408358693122864, -0.04037027806043625, 0.07750772684812546, 0.04085922613739967, 0.02690250426530838, -0.012930345721542835, -0.09768964350223541, 0.03800887614488602, -0.09662877023220062, 0.054791923612356186, -0.06991695612668991, -0.1949336975812912, -0.061280205845832825, 0.0023225415498018265, 0.039309412240982056, 0.022137563675642014, 0.10605164617300034, 0.017101019620895386, -0.03633737191557884, 0.05713365226984024, 0.0005838759243488312, -0.01225859671831131, 0.0020293663255870342, -0.01706792414188385, 0.11154720187187195, 0.007104562595486641, 0.021614383906126022, -0.12169289588928223, 0.053943172097206116, -0.05383014306426048, 0.018748503178358078, 0.029469531029462814, -0.06176181882619858, 0.01860296167433262, 0.024166876450181007, -0.0015528658404946327, -0.13261441886425018, -0.16401541233062744, 0.02581726387143135, 0.007949343882501125, -0.08364351093769073, -0.10428491234779358, -0.09974739700555801, -0.05713500455021858, 0.01597749814391136, -0.019052930176258087, 0.012956537306308746, -0.043037742376327515, 0.07441737502813339, -0.03828577697277069, 0.1370566189289093, -0.09535866975784302, 0.05127450078725815, -0.10738140344619751, -0.04108767956495285, -0.0804896354675293, 0.08953788876533508, -0.0030834178905934095, 0.04695384204387665, 0.011799615807831287, -0.03455916792154312, -0.04610372707247734, 0.037143774330616, -0.0524589940905571, 0.1780407726764679, -0.1465904265642166, -0.08785836398601532, 0.27364280819892883, -0.10529017448425293, -0.1866082400083542, 0.14749449491500854, -0.009997260756790638, 0.054774001240730286, 0.1019631177186966, 0.20435383915901184, 0.0019123286474496126, 0.023100290447473526, 0.10274538397789001, 0.08608831465244293, -0.07944673299789429, 0.008127938956022263, 0.03658723831176758, 0.015190929174423218, -0.05177658796310425, 0.03792482614517212, 0.00911007635295391, 0.07040978223085403, -0.006703962571918964, -0.016700847074389458, 0.0037047970108687878, 0.008876058273017406, 0.05781126022338867, -0.028141673654317856, 0.13755415380001068, -0.059995852410793304, -0.025661583989858627, -0.10463643819093704, 0.0002251066907774657, -0.04127488285303116, 0.09056096524000168, -0.00512147881090641, 0.1049070954322815, 0.019593439996242523, 0.0791654884815216, -0.12508417665958405, -0.010133717209100723, -0.03017987310886383, 0.1124357134103775, 0.06284242868423462, 0.09446145594120026, 0.06871743500232697, 0.02889222279191017, -0.049836866557598114, 0.06823620200157166, 0.16815853118896484, -0.03449784591794014, -0.08621354401111603, -0.13682137429714203, 0.06752511858940125, -0.0629672259092331, 0.06575429439544678, -0.10528282821178436, 0.041321106255054474, 0.016233770176768303, 0.09273680299520493, -0.028000373393297195, 0.01569272018969059, 0.0009924957994371653, -0.022767841815948486, -0.08932660520076752, -0.03777209669351578, 0.0933055430650711, 0.0053114378824830055, -0.09579470008611679, 0.23412001132965088, -0.18661560118198395, 0.11514696478843689, 0.18844285607337952, -0.23562726378440857, 0.004645537585020065, -0.08615560829639435, -0.032945163547992706, -0.012463530525565147, 0.06846268475055695, -0.008126414380967617, 0.20710298418998718, 0.02152816578745842, 0.1799556016921997, -0.03299377113580704, -0.055573053658008575, -0.06548352539539337, -0.03873597830533981, 0.013585329055786133, 0.12083686888217926, 0.10954977571964264, -0.09471899271011353, 0.16591253876686096, 0.06281638145446777, 0.01285959780216217, 0.1989755481481552, 0.10162375867366791, 0.018465086817741394, 0.05261732265353203, -0.009915602393448353, -0.016312487423419952, -0.06617686152458191, -0.2720363140106201, -0.022636009380221367, 0.09089183807373047, -0.000044819898903369904, 0.08634502440690994, -0.1177721917629242, -0.05302949622273445, 0.0027246414683759212, -0.009751816280186176, 0.05398271232843399, 0.1520978808403015, -0.020374011248350143, 0.12683804333209991, -0.03150931000709534, -0.07125429064035416, 0.06139340251684189, 0.007860809564590454, -0.06915654987096786, 0.17656075954437256, -0.13541632890701294, -0.30423009395599365, -0.10002188384532928, -0.20794804394245148, -0.08806388080120087, 0.05021306127309799, 0.08505034446716309, -0.1202472522854805, -0.014454866759479046, -0.013682014308869839, 0.1025705486536026, -0.1784762293100357, -0.027358150109648705, -0.07787293195724487, -0.004314797930419445, -0.17612019181251526, -0.09081652015447617, -0.03819951042532921, -0.014087172225117683, -0.0841885656118393, 0.14302858710289001, -0.13169878721237183, 0.002548591000959277, 0.213802307844162, 0.027100462466478348, 0.06824927777051926, -0.04536136984825134, 0.2325112521648407, -0.08249329030513763, 0.021983947604894638, 0.10190591961145401, -0.011717324145138264, 0.0727166086435318, 0.1207466870546341, 0.023063041269779205, -0.06990478932857513, 0.002824747934937477, -0.032633766531944275, -0.08337754011154175, -0.19988340139389038, -0.06400222331285477, -0.1320084035396576, 0.15760111808776855, 0.06374546885490417, 0.08850138634443283, 0.15267208218574524, 0.07417335361242294, -0.011941826902329922, 0.027392378076910973, 0.09842611849308014, 0.10804754495620728, 0.26540398597717285, -0.04679965227842331, 0.0976310446858406, 0.011697877198457718, -0.12767644226551056, 0.06473664939403534, 0.11575727164745331, 0.10438727587461472, 0.058607831597328186, 0.0678510069847107, -0.020579097792506218, 0.01497538574039936, 0.10904894769191742, 0.03608834743499756, 0.04561825096607208, -0.025500640273094177, -0.0418923981487751, -0.05734262987971306, 0.008914070203900337, 0.07560605555772781, 0.019149649888277054, -0.11067512631416321, 0.00275624543428421, 0.04348788782954216, 0.05517588183283806, -0.020587924867868423, 0.07209046185016632, -0.13969957828521729, -0.025503646582365036, 0.0856671929359436, -0.011570225469768047, -0.11118893325328827, 0.12472975254058838, 0.039318010210990906, -0.15119268000125885, 0.07064856588840485, -0.05994023010134697, 0.09893705695867538, -0.030644794926047325, 0.0679929107427597, -0.08408012241125107, -0.10547026246786118, 0.0003558758180588484, 0.08076029270887375, -0.2821684777736664, 0.17951267957687378, -0.03808591514825821, -0.03410544991493225, -0.04985491558909416, -0.021032677963376045, 0.056860048323869705, 0.11849650740623474, 0.12979209423065186, -0.0019196383655071259, 0.07829723507165909, 0.01881568692624569, -0.00170033797621727, 0.016322525218129158, 0.046422455459833145, -0.055369578301906586, -0.0158549714833498, -0.02669402025640011, 0.0019921984057873487, -0.048932477831840515, -0.033079806715250015, 0.04292672872543335, -0.22484149038791656, 0.09117942303419113, 0.10809938609600067, 0.028686555102467537, 0.033663295209407806, -0.06156184524297714, -0.08281350135803223, 0.22607377171516418, 0.014403248205780983, -0.1295873522758484, -0.06308655440807343, 0.032366979867219925, 0.059771209955215454, -0.0355374850332737, -0.018700862303376198, -0.09346245229244232, 0.04255778715014458, -0.1265409141778946, -0.142607644200325, 0.08159507811069489, -0.05655594915151596, -0.06316933780908585, -0.027429254725575447, 0.1880079060792923, -0.02114991471171379, 0.07580824196338654, 0.0037714382633566856, 0.03305105119943619, -0.13178177177906036, -0.09944485127925873, -0.014336314052343369, -0.054112762212753296, -0.05475674569606781, 0.008005809970200062, 0.04339006543159485, 0.0872558206319809, -0.07871737331151962, -0.028180301189422607, 0.3180432915687561, 0.14157041907310486, -0.041152141988277435, 0.16018512845039368, 0.06227409839630127, -0.06818420439958572, -0.23930665850639343, -0.1138848066329956, -0.10739355534315109, -0.0559304840862751, -0.059126634150743484, -0.14686015248298645, 0.09206750988960266, -0.05726713687181473, 0.0061450679786503315, 0.15972265601158142, -0.3005811870098114, -0.11088323593139648, 0.14024609327316284, 0.017290394753217697, 0.3640836775302887, -0.1228390634059906, -0.055082254111766815, -0.015156857669353485, -0.16496481001377106, 0.09070722758769989, 0.015410330146551132, 0.11901314556598663, -0.0836920216679573, 0.18978270888328552, 0.03005683794617653, -0.0068378145806491375, 0.1130145937204361, 0.07239632308483124, -0.04066509008407593, -0.08782052248716354, -0.0776001363992691, 0.01842515356838703, 0.037652015686035156, 0.013590807095170021, -0.06284631043672562, -0.004790465347468853, -0.08503337949514389, -0.03819546848535538, -0.09164988249540329, -0.0015074126422405243, -0.0014627352356910706, -0.057050637900829315, -0.021886419504880905, -0.0017445459961891174, -0.0035965461283922195, 0.03909335657954216, 0.1703273355960846, -0.04901299625635147, 0.1574823409318924, 0.0260554738342762, 0.07109086215496063, -0.13442687690258026, -0.060555048286914825, -0.03622489050030708, -0.025915447622537613, 0.06509897112846375, -0.06936880201101303, -0.030921362340450287, 0.15102329850196838, -0.020268825814127922, 0.045350465923547745, 0.1292015016078949, 0.013731560669839382, 0.0053041852079331875, 0.0570853166282177, -0.24560818076133728, -0.15736152231693268, -0.052636634558439255, -0.05197368562221527, 0.1274430900812149, 0.04840640723705292, 0.19958584010601044, -0.06594612449407578, -0.03397516533732414, 0.05736926943063736, 0.013749856501817703, -0.0017229977529495955, 0.07987318933010101, -0.015440300107002258, 0.006659801583737135, -0.15620073676109314, 0.07459729164838791, 0.006896140519529581, -0.10052015632390976, 0.03169224038720131, 0.18870747089385986, -0.12493041157722473, -0.10614178329706192, -0.11547666788101196, 0.07271894812583923, -0.10056423395872116, 0.0002751704305410385, -0.0046763671562075615, -0.1319015920162201, 0.05779172480106354, -0.00038787536323070526, 0.061845242977142334, 0.05237482488155365, -0.15944844484329224, -0.03268345072865486, -0.02349804900586605, 0.042672306299209595, 0.047068655490875244, 0.006903572473675013, -0.058427393436431885, 0.0994153842329979, -0.05051529407501221, 0.09632633626461029, -0.06777458637952805, -0.1200634092092514, -0.0927400290966034, 0.03964018076658249, -0.13151314854621887, -0.0692918598651886, -0.11806371062994003, -0.04273033142089844, -0.004625196103006601, -0.035258129239082336, -0.032276324927806854, -0.031054425984621048, -0.11201182752847672, 0.003935540094971657, -0.0470658615231514, -0.011552292853593826, -0.06612269580364227, -0.003940910566598177, 0.05805809050798416, -0.03935599699616432, 0.13867168128490448, 0.16647858917713165, -0.15514089167118073, 0.07677218317985535, -0.1685427725315094, -0.07516123354434967, 0.06984034180641174, 0.03165632486343384, 0.05344364419579506, 0.059440046548843384, -0.025441037490963936, 0.017906736582517624, 0.03498870134353638, 0.07840894162654877, 0.053051963448524475, -0.07841214537620544, 0.030624940991401672, -0.05290471762418747, -0.12006967514753342, -0.000231156125664711, -0.0226430706679821, 0.007653330452740192, 0.03761737421154976, 0.06373873353004456, -0.0760762169957161, 0.08361746370792389, -0.0737353041768074, 0.04220488294959068, 0.028027689084410667, -0.15039730072021484, -0.00910397619009018, -0.07054219394922256, 0.02919093519449234, -0.021660130470991135, 0.19487307965755463, -0.01624002307653427, 0.00781339779496193, 0.018008975312113762, 0.07349169254302979, 0.01488073542714119, -0.021530229598283768, 0.1621304750442505, 0.11021947860717773, -0.08122758567333221, -0.04515838623046875, 0.08057281374931335, 0.07687249779701233, 0.07451280206441879, 0.1162770465016365, -0.07681170105934143, -0.03407645598053932, 0.07622746378183365, -0.039603374898433685, 0.06729553639888763, -0.09732499718666077, -0.10669039189815521, 0.04559433460235596, 0.05374941602349281, -0.0527818463742733, 0.16816911101341248, 0.17918722331523895, 0.013131311163306236, 0.012708173133432865, -0.01733400672674179, -0.07395055890083313, -0.1567104011774063, -0.1784777045249939, -0.05647404119372368, -0.11609060317277908, 0.005330994725227356, -0.14184488356113434, 0.04775647073984146, -0.0017415538895875216, 0.12955868244171143, -0.06126172095537186, 0.07519124448299408, 0.09107817709445953, -0.08536062389612198, 0.0806393101811409, -0.03553549572825432, 0.09202031046152115, -0.011557330377399921, 0.044254809617996216, -0.08543217927217484, -0.008890951052308083, -0.006461475044488907, 0.06062658876180649, -0.02894883044064045, 0.01660233736038208, -0.13762788474559784, -0.07014933973550797, -0.059105195105075836, 0.07127918303012848, -0.029436225071549416, 0.1610472947359085, 0.05126805603504181, -0.04614876210689545, 0.014419383369386196, 0.1970549076795578, -0.052382491528987885, -0.09264914691448212, -0.031681641936302185, 0.16575223207473755, 0.021530043333768845, 0.05801204964518547, -0.01118925865739584, 0.02923089638352394, -0.11139746755361557, 0.27545517683029175, 0.3634028136730194, -0.11353981494903564, -0.012144984677433968, 0.02952711284160614, 0.04699595272541046, 0.12081082165241241, 0.047329820692539215, 0.09195783734321594, 0.3007553517818451, -0.09227307140827179, -0.0475301519036293, -0.056458525359630585, -0.026054421439766884, -0.07386115193367004, 0.05792106315493584, 0.10494376718997955, -0.059347741305828094, -0.0452754832804203, 0.08314090967178345, -0.252031147480011, 0.09283767640590668, -0.16627268493175507, -0.18303143978118896, -0.05790427699685097, 0.010824799537658691, 0.028273558244109154, 0.0028769182972609997, 0.08227575570344925, 0.01287081465125084, -0.07710790634155273, 0.10914352536201477, 0.010229930281639099, -0.22411158680915833, 0.028085123747587204, 0.10111425817012787, -0.0828917846083641, 0.001978249754756689, -0.037217266857624054, 0.048161014914512634, 0.0860811322927475, 0.07358460128307343, -0.014528268948197365, -0.015057718381285667, 0.000730111263692379, 0.040068119764328, -0.008920635096728802, 0.0557340569794178, 0.04994771629571915, -0.1317262351512909, 0.09821461141109467, -0.060323528945446014, 0.03660005331039429, 0.06039358675479889, -0.007528811693191528, -0.020978445187211037, 0.020991196855902672, -0.042133674025535583, 0.03819144517183304, 0.11411964893341064, -0.005739775486290455, 0.00021349964663386345, -0.03667907044291496, -0.07916922122240067, -0.006765095517039299, -0.027232836931943893, -0.06285584717988968, -0.13698558509349823, -0.1051611453294754, 0.0030524462927132845, -0.010489038191735744, -0.20800088346004486, -0.021541273221373558, -0.1378927230834961, 0.04871266335248947, -0.09494157880544662, 0.10229058563709259, 0.051117390394210815, 0.02578243613243103, -0.009486589580774307, -0.008693112060427666, 0.038259491324424744, 0.09200619906187057, -0.1385052502155304, -0.08123001456260681 ]
null
null
transformers
# Harry Potter DialoGPT small Model (Movies 1 through 3)
{"tags": ["conversational"]}
text-generation
Ayran/DialoGPT-small-harry-potter-1-through-3
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Harry Potter DialoGPT small Model (Movies 1 through 3)
[ "# Harry Potter DialoGPT small Model (Movies 1 through 3)" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Harry Potter DialoGPT small Model (Movies 1 through 3)" ]
[ 51, 15 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Harry Potter DialoGPT small Model (Movies 1 through 3)" ]
[ -0.001265437458641827, 0.05945855379104614, -0.005430597811937332, 0.08816283941268921, 0.1294340044260025, 0.052468813955783844, 0.21999667584896088, 0.13915419578552246, 0.002475939691066742, -0.05605195835232735, 0.07804138958454132, 0.16504253447055817, 0.022258611395955086, 0.048143789172172546, -0.03527703881263733, -0.33563584089279175, 0.006698964163661003, 0.012024195864796638, -0.05030691996216774, 0.09974672645330429, 0.10812358558177948, -0.06090869382023811, 0.08307799696922302, -0.015465198084712029, -0.1340479999780655, -0.050895772874355316, 0.04432493448257446, -0.10104735195636749, 0.16193287074565887, 0.04047083109617233, -0.003194481134414673, -0.016873203217983246, -0.012488776817917824, -0.09670485556125641, 0.03543306887149811, -0.05548941716551781, 0.02967393770813942, 0.015892544761300087, 0.07159064710140228, -0.0007175600039772689, 0.179011270403862, 0.17587588727474213, 0.0899214819073677, -0.0000386889478249941, -0.09712957590818405, -0.09413819760084152, 0.007486782968044281, 0.03318898379802704, -0.050949882715940475, 0.08117260783910751, -0.010072079487144947, 0.10480542480945587, -0.09115993976593018, 0.07302570343017578, 0.1782027631998062, -0.3518826365470886, -0.05009450018405914, 0.10228756815195084, 0.11942420899868011, 0.12337330728769302, -0.15438632667064667, 0.02503092586994171, -0.028547175228595734, 0.00795923825353384, -0.020604100078344345, -0.08399597555398941, -0.07330860942602158, -0.013460521586239338, -0.13655021786689758, 0.011272147297859192, 0.22599227726459503, -0.10343845933675766, 0.007739400025457144, -0.12751957774162292, 0.0008588816272094846, 0.022688886150717735, -0.08565889298915863, -0.06084126606583595, -0.02481348253786564, 0.0210021510720253, -0.0252126045525074, -0.03885525465011597, -0.08230645209550858, -0.001815312192775309, -0.05482272431254387, 0.14740367233753204, 0.049647778272628784, 0.06280142068862915, -0.2563231587409973, 0.06212084740400314, 0.043718237429857254, -0.02556716836988926, 0.02741410955786705, -0.12331018596887589, 0.11894502490758896, 0.007510469760745764, -0.034534502774477005, 0.03651842847466469, 0.021916907280683517, 0.14117331802845, 0.012235518544912338, 0.0091316569596529, 0.013168100267648697, 0.08401157706975937, 0.08968181908130646, 0.03956500440835953, 0.015532314777374268, -0.09412485361099243, 0.06991294771432877, -0.00944199226796627, 0.03598126024007797, -0.018362756818532944, -0.1928238421678543, -0.049810271710157394, 0.04166112467646599, 0.019937369972467422, 0.060761671513319016, 0.05710085481405258, -0.01762264408171177, -0.05382545292377472, 0.0452469140291214, 0.02744658850133419, -0.03010224923491478, -0.011312478221952915, -0.024765942245721817, 0.25112423300743103, -0.03662501648068428, -0.011384794488549232, -0.08709299564361572, 0.10676499456167221, -0.06790044903755188, 0.016811316832900047, 0.046495821326971054, -0.03790568560361862, 0.02642630785703659, 0.04230424389243126, 0.04013695567846298, -0.13205456733703613, -0.057782188057899475, 0.017492657527327538, -0.021364912390708923, -0.05763585865497589, -0.10249415785074234, -0.05275065451860428, 0.006106220185756683, 0.06119120493531227, -0.0036619342863559723, -0.020298169925808907, -0.05589964985847473, 0.062201812863349915, -0.02055038884282112, 0.12308355420827866, -0.057958249002695084, 0.040217943489551544, -0.052025265991687775, -0.07412064075469971, -0.13537929952144623, 0.02507772669196129, 0.024700982496142387, 0.026166187599301338, 0.04966914281249046, -0.05162941291928291, 0.03643125668168068, 0.05808240920305252, -0.12608666718006134, 0.19757398962974548, -0.08413363248109818, -0.1269601285457611, 0.20151418447494507, -0.10064554959535599, -0.18279409408569336, 0.14727044105529785, 0.012795780785381794, 0.061719030141830444, 0.07158979028463364, 0.22349517047405243, 0.0018105664057657123, -0.0376456156373024, 0.06902137398719788, 0.1339782476425171, -0.07936478406190872, 0.05037826672196388, 0.07797706872224808, 0.044852614402770996, -0.04083162546157837, 0.03960878401994705, -0.025857584550976753, -0.0279153473675251, -0.044727738946676254, 0.010585769079625607, 0.04119398072361946, -0.018498815596103668, 0.1365087926387787, -0.030991384759545326, 0.1494596004486084, -0.07981754839420319, -0.01667008176445961, -0.029736321419477463, 0.09151162207126617, -0.012432990595698357, 0.1180863231420517, -0.05499103292822838, 0.12536606192588806, 0.09611207246780396, 0.02393028698861599, -0.0949275866150856, 0.020379433408379555, -0.01994355581700802, 0.17065851390361786, 0.045232634991407394, 0.07309222221374512, 0.054651275277137756, 0.025557998567819595, -0.0917103961110115, 0.08682231605052948, 0.09532257169485092, -0.021423963829874992, -0.05292654037475586, -0.2062467336654663, 0.004499196074903011, -0.06368251889944077, 0.12360783666372299, -0.10078094899654388, -0.009849398396909237, -0.04514830932021141, 0.05216563493013382, -0.027143431827425957, -0.014406410045921803, 0.012943938374519348, -0.03689597174525261, -0.060288846492767334, -0.04082426428794861, 0.09602398425340652, -0.0064531913958489895, 0.0020828458946198225, 0.17044198513031006, -0.1509871482849121, 0.1628507822751999, 0.19157736003398895, -0.28570985794067383, -0.015681182965636253, -0.0980018824338913, -0.041473135352134705, 0.04558561369776726, 0.056391846388578415, 0.02105051279067993, 0.13745829463005066, -0.01782030053436756, 0.18088099360466003, -0.03630528971552849, -0.05330076813697815, -0.061569131910800934, -0.07193028181791306, -0.022512955591082573, 0.08315333724021912, 0.06472110748291016, -0.061331856995821, 0.11658936738967896, 0.09959034621715546, 0.08565603196620941, 0.171228289604187, 0.10675961524248123, 0.0037812585942447186, 0.09288536012172699, -0.06524651497602463, -0.003539694007486105, -0.06658078730106354, -0.268503338098526, -0.06958892196416855, 0.09874369204044342, -0.03327863663434982, 0.06044897437095642, -0.08953047543764114, -0.058938127011060715, 0.012139864265918732, 0.008878803811967373, 0.043798740953207016, 0.11329653114080429, -0.014212112873792648, 0.1663285493850708, 0.0366402193903923, -0.0574771948158741, 0.07197699695825577, -0.0020418751519173384, -0.1048269122838974, 0.14710268378257751, -0.13430319726467133, -0.28486648201942444, -0.059946175664663315, -0.19331517815589905, 0.04820682480931282, 0.053364239633083344, 0.05694647133350372, -0.10906992852687836, -0.008411417715251446, -0.0067405253648757935, 0.01516283955425024, -0.29022637009620667, -0.026822609826922417, -0.08989319950342178, 0.07651493698358536, -0.19939875602722168, -0.10167872905731201, -0.013883950188755989, -0.0028510878328233957, -0.09725581854581833, 0.13509473204612732, -0.13182836771011353, -0.04011337831616402, 0.21293306350708008, 0.006737926509231329, 0.046530451625585556, -0.07255145162343979, 0.19988486170768738, -0.08581612259149551, -0.028682442381978035, 0.1710389107465744, -0.04097841680049896, 0.02527467906475067, 0.05544913187623024, 0.0070506208576262, -0.05809053033590317, 0.004132918547838926, -0.09937922656536102, -0.0247059166431427, -0.21795998513698578, -0.024484366178512573, -0.09891988337039948, 0.16057328879833221, 0.06746519356966019, 0.05139734223484993, 0.12732352316379547, 0.042880699038505554, -0.020583340898156166, -0.018920153379440308, 0.10413850843906403, 0.1317727118730545, 0.22296728193759918, -0.015064418315887451, 0.08490486443042755, 0.004824802745133638, -0.03174498304724693, 0.07869286835193634, -0.0016078522894531488, 0.0009414987871423364, 0.02637101337313652, 0.01468638051301241, -0.03432895615696907, 0.05589477717876434, 0.13453428447246552, 0.021247997879981995, 0.07722742855548859, -0.0018379980465397239, -0.07403557747602463, -0.014178807847201824, -0.04786340147256851, 0.05289877578616142, 0.037598446011543274, -0.16577370464801788, -0.10336516797542572, 0.07769923657178879, 0.03548917919397354, -0.02796231023967266, 0.08645686507225037, -0.13586926460266113, -0.03252458944916725, 0.06291027367115021, -0.049619950354099274, -0.07702522724866867, 0.10497265309095383, 0.003003255696967244, -0.25528350472450256, 0.040965985506772995, 0.03906792029738426, 0.060349833220243454, -0.08530276268720627, 0.039376772940158844, -0.10880283266305923, -0.14007051289081573, 0.017584608867764473, 0.07037345319986343, -0.30727556347846985, 0.08393405377864838, -0.028895098716020584, -0.009619630873203278, -0.039726197719573975, -0.04631941765546799, 0.0324396938085556, 0.08477717638015747, 0.12497278302907944, 0.020565636456012726, 0.12042811512947083, 0.022883739322423935, 0.04665498062968254, -0.00861417781561613, 0.11868876218795776, -0.013684642501175404, 0.011609159409999847, -0.09376613050699234, -0.02087409235537052, -0.07653187215328217, -0.06444594264030457, 0.10567328333854675, -0.18348285555839539, 0.10333292186260223, -0.024428674951195717, 0.033187177032232285, 0.01489071175456047, -0.06629665940999985, -0.07264319062232971, 0.16234862804412842, 0.004240548703819513, -0.11776222288608551, -0.0688256174325943, -0.018566643819212914, 0.06137774884700775, -0.0013119467766955495, 0.0559600293636322, -0.047369975596666336, 0.0806114673614502, -0.14782670140266418, -0.1526302844285965, 0.07882741838693619, -0.09122192859649658, -0.14629118144512177, -0.028300834819674492, 0.18957936763763428, -0.06712496280670166, 0.11339187622070312, 0.01594792678952217, -0.014724833890795708, -0.13712000846862793, -0.013733158819377422, 0.006175227463245392, 0.010088182985782623, 0.0029911829624325037, 0.0039472561329603195, 0.02107408083975315, 0.07936922460794449, -0.0767984539270401, -0.04569636657834053, 0.31585440039634705, 0.23638640344142914, -0.021758083254098892, 0.09220116585493088, 0.07815555483102798, -0.05452229827642441, -0.17964045703411102, -0.11954894661903381, -0.08184997737407684, -0.12887577712535858, -0.06526225060224533, -0.12513908743858337, 0.09808563441038132, -0.04942462593317032, 0.02263471484184265, 0.15399718284606934, -0.31418466567993164, -0.1152205839753151, 0.09881803393363953, 0.04279426857829094, 0.37830668687820435, -0.11069262027740479, -0.000770594400819391, -0.0428449772298336, -0.14475956559181213, 0.04988851025700569, 0.018015926703810692, 0.13332591950893402, -0.09380822628736496, 0.1494360715150833, -0.026775291189551353, 0.012053807266056538, 0.07397574186325073, -0.006433872506022453, -0.09055628627538681, -0.071780264377594, -0.18736766278743744, -0.01315106451511383, 0.022885825484991074, -0.006525689270347357, 0.005030883941799402, -0.019141271710395813, -0.09101752191781998, -0.04529952257871628, -0.09000089019536972, -0.02403157763183117, -0.002859959378838539, -0.04427497088909149, -0.08670295774936676, 0.020152969285845757, -0.05871886387467384, 0.0180633794516325, 0.0952385887503624, -0.10049955546855927, 0.13309445977210999, -0.018664345145225525, 0.1469445377588272, -0.09394396841526031, -0.09643745422363281, -0.03954927250742912, -0.0396239347755909, 0.020716890692710876, -0.015125975012779236, -0.03924591839313507, 0.12669037282466888, -0.005019704811275005, 0.04485461860895157, 0.1272626370191574, 0.06419500708580017, 0.03193134441971779, 0.06886524707078934, -0.18466512858867645, -0.09364386647939682, -0.047386590391397476, 0.036720674484968185, 0.030958397313952446, -0.017972683534026146, 0.15989172458648682, -0.07032763957977295, -0.10478522628545761, -0.015489395707845688, 0.02505795657634735, -0.008929071016609669, 0.0740518718957901, 0.02314537577331066, 0.022179603576660156, -0.16522099077701569, 0.11224374920129776, -0.018886364996433258, -0.0735638216137886, 0.01483741495758295, 0.2575080394744873, -0.1206425279378891, -0.08391071856021881, -0.04239973425865173, 0.026805981993675232, -0.07985082268714905, 0.019581425935029984, -0.01803322695195675, -0.14790867269039154, 0.056206442415714264, 0.02812010608613491, 0.036761630326509476, 0.07380270957946777, -0.13085585832595825, -0.0016986601985991001, -0.05551668629050255, 0.009154348634183407, 0.041016872972249985, 0.02369966357946396, -0.09346571564674377, 0.17384101450443268, -0.08426673710346222, 0.09288346022367477, -0.07358769327402115, -0.09536053985357285, -0.10711664706468582, 0.029116494581103325, -0.06389486789703369, -0.028238626196980476, -0.1421726793050766, -0.04623403400182724, -0.017494726926088333, 0.002449760679155588, -0.022908281534910202, -0.042479950934648514, -0.11104658991098404, -0.0022088142577558756, -0.04053768143057823, -0.03431041166186333, -0.02337193861603737, 0.03755253553390503, 0.06421105563640594, -0.05876101925969124, 0.14884981513023376, 0.20260857045650482, -0.08839020133018494, 0.09004741907119751, -0.06474936008453369, -0.18695664405822754, 0.029920540750026703, -0.013208777643740177, 0.012628631666302681, 0.08365584909915924, -0.02990761771798134, 0.004299277905374765, 0.03760553151369095, 0.07606223225593567, 0.0764947459101677, -0.057286426424980164, 0.025400223210453987, -0.04002956673502922, -0.11804735660552979, -0.007573253940790892, -0.03554921969771385, 0.08943560719490051, -0.04314231872558594, 0.09322191774845123, -0.08548704534769058, 0.05833054333925247, -0.08720943331718445, 0.037755414843559265, -0.011896003969013691, -0.12425757944583893, -0.020102253183722496, -0.059266164898872375, 0.03303524851799011, 0.01252033468335867, 0.15848010778427124, -0.015480940230190754, 0.006128361448645592, 0.033800940960645676, 0.04072956740856171, 0.013624565675854683, 0.005457428749650717, 0.19054602086544037, 0.13696233928203583, -0.10136112570762634, -0.04476577043533325, 0.054778508841991425, 0.11172035336494446, 0.1200665608048439, 0.1318596601486206, 0.016045168042182922, -0.03632000461220741, 0.07917598634958267, -0.08634895831346512, 0.06269717216491699, -0.11731816828250885, -0.11923974007368088, 0.0768299475312233, -0.0008189158979803324, -0.08527418226003647, 0.1705496460199356, 0.22740308940410614, 0.014274323359131813, 0.02191457711160183, -0.04937898740172386, -0.08676698803901672, -0.1503264456987381, -0.035867370665073395, -0.05881569907069206, -0.14976175129413605, -0.009336358867585659, -0.1611025333404541, 0.05384054034948349, 0.0009340848191641271, 0.15010559558868408, -0.09009499102830887, 0.008950614370405674, 0.10187651962041855, -0.11291585862636566, 0.12715652585029602, -0.01931251399219036, 0.09020444005727768, -0.014630625955760479, -0.010354593396186829, -0.10052064061164856, -0.03781905025243759, 0.06222230941057205, 0.0722304955124855, -0.12224172800779343, 0.015795139595866203, -0.09554126858711243, -0.10458918660879135, -0.0155441639944911, 0.057473570108413696, -0.030075982213020325, 0.16920198500156403, 0.0069958362728357315, -0.046418335288763046, -0.0033595971763134003, 0.25930434465408325, -0.10158295929431915, -0.11017567664384842, 0.00805196724832058, 0.1942020058631897, -0.011687560938298702, 0.06194041296839714, -0.03125372156500816, 0.03166089951992035, -0.08654371649026871, 0.28872305154800415, 0.36201801896095276, -0.15356552600860596, -0.030987443402409554, 0.04133684188127518, 0.04551887512207031, 0.07470954209566116, 0.09515615552663803, 0.08018577843904495, 0.27063441276550293, -0.1208561509847641, 0.030451154336333275, -0.01913062110543251, 0.027534643188118935, -0.043131861835718155, 0.01846606843173504, 0.10828310996294022, -0.044134873896837234, -0.06420867890119553, 0.04933067038655281, -0.24877937138080597, 0.12614406645298004, -0.13090483844280243, -0.13803012669086456, -0.012324601411819458, 0.0033851086627691984, -0.014778392389416695, 0.036149006336927414, 0.1176581084728241, 0.03895815461874008, -0.0628524050116539, 0.005778856575489044, 0.02855873852968216, -0.23354284465312958, -0.032797422260046005, 0.17819300293922424, -0.035236895084381104, 0.03170764073729515, -0.05455877631902695, 0.06619163602590561, 0.024358732625842094, 0.0851525068283081, -0.06129913404583931, -0.10046616941690445, -0.010776383802294731, -0.01874607428908348, -0.0002868062583729625, 0.12941573560237885, 0.010270650498569012, -0.0755012258887291, 0.1124432310461998, -0.04366844519972801, 0.05522306263446808, 0.06402573734521866, -0.00462582940235734, 0.009246902540326118, -0.017647098749876022, -0.07962635904550552, 0.06183365359902382, 0.13192924857139587, -0.03698975592851639, -0.00961978267878294, -0.02431817539036274, -0.04400881379842758, 0.0400070995092392, -0.013891751877963543, -0.06020725145936012, -0.18202130496501923, -0.12881827354431152, -0.017382891848683357, -0.005727801006287336, -0.19525206089019775, 0.017865175381302834, -0.1251579076051712, 0.04065266251564026, -0.14150883257389069, 0.07313301414251328, 0.04909614101052284, 0.020075371488928795, 0.007590079214423895, 0.04539278522133827, 0.0655611902475357, 0.14945626258850098, -0.16561062633991241, -0.04529406130313873 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ai-club-inductions-21-nlp-ALBERT This model was trained from scratch on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cpu - Datasets 1.14.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "ai-club-inductions-21-nlp-ALBERT", "results": []}]}
question-answering
AyushPJ/ai-club-inductions-21-nlp-ALBERT
[ "transformers", "pytorch", "albert", "question-answering", "generated_from_trainer", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #albert #question-answering #generated_from_trainer #endpoints_compatible #region-us
# ai-club-inductions-21-nlp-ALBERT This model was trained from scratch on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cpu - Datasets 1.14.0 - Tokenizers 0.10.3
[ "# ai-club-inductions-21-nlp-ALBERT\n\nThis model was trained from scratch on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #albert #question-answering #generated_from_trainer #endpoints_compatible #region-us \n", "# ai-club-inductions-21-nlp-ALBERT\n\nThis model was trained from scratch on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ 37, 31, 6, 12, 8, 3, 90, 33 ]
[ "passage: TAGS\n#transformers #pytorch #albert #question-answering #generated_from_trainer #endpoints_compatible #region-us \n# ai-club-inductions-21-nlp-ALBERT\n\nThis model was trained from scratch on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ -0.09546796977519989, 0.07245520502328873, -0.001882903859950602, 0.097852922976017, 0.1316424310207367, -0.012304051779210567, 0.13833926618099213, 0.12314573675394058, -0.08174196630716324, 0.05798599123954773, 0.12029539793729782, 0.09692655503749847, 0.03489597141742706, 0.14911869168281555, -0.03965099900960922, -0.23490780591964722, 0.03643157705664635, 0.07349773496389389, 0.010279469192028046, 0.11739026755094528, 0.0861370787024498, -0.14772076904773712, 0.0888369157910347, 0.042861245572566986, -0.1966889649629593, 0.00950834434479475, -0.017015719786286354, -0.09759540110826492, 0.09887423366308212, -0.009890522807836533, 0.11127837002277374, 0.012986334972083569, 0.09057465195655823, -0.13623890280723572, 0.005834309384226799, 0.03954274579882622, 0.03641489893198013, 0.10770758241415024, 0.03598989546298981, -0.010917012579739094, 0.06288689374923706, -0.10888975113630295, 0.10136407613754272, 0.029890725389122963, -0.11633162200450897, -0.22886551916599274, -0.09204602986574173, 0.08050847798585892, 0.0634346604347229, 0.09175005555152893, -0.0023719463497400284, 0.2300216406583786, -0.06110342964529991, 0.08283645659685135, 0.29738515615463257, -0.33010125160217285, -0.061282992362976074, 0.01797262206673622, 0.07200857251882553, 0.050852514803409576, -0.08854319900274277, -0.023147178813815117, 0.06548231840133667, 0.06346306204795837, 0.1472795307636261, -0.026607278734445572, -0.02893718145787716, -0.013835257850587368, -0.14103829860687256, -0.04648822546005249, 0.15688277781009674, 0.025989534333348274, -0.07031938433647156, -0.07539839297533035, -0.07389391213655472, -0.13694290816783905, -0.038971491158008575, -0.0605238676071167, 0.03782428428530693, -0.07387550920248032, -0.07742322236299515, -0.02052326500415802, -0.051915448158979416, -0.040741290897130966, -0.04155896604061127, 0.18182919919490814, 0.05075938254594803, 0.05126788839697838, -0.0709245428442955, 0.0828840509057045, -0.044701214879751205, -0.12405695766210556, -0.0188333448022604, -0.01382120605558157, -0.004214324522763491, -0.054155442863702774, -0.03705647215247154, -0.03706659749150276, -0.025543611496686935, 0.17029397189617157, -0.0874912291765213, 0.04120801016688347, 0.017063530161976814, 0.020023247227072716, -0.014105568639934063, 0.19104507565498352, -0.0552438348531723, 0.014838293194770813, 0.016246534883975983, 0.07182956486940384, 0.019450882449746132, -0.022936027497053146, -0.1099461168050766, -0.004984461236745119, 0.12011837959289551, 0.02631060965359211, -0.06590654700994492, 0.04567011818289757, -0.029322607442736626, -0.055101487785577774, 0.03094988316297531, -0.11637350171804428, 0.04266015440225601, -0.02365671657025814, -0.09880252927541733, 0.02396765723824501, 0.006327996030449867, -0.015362088568508625, -0.04715530574321747, 0.05569799244403839, -0.10069039463996887, 0.03657347708940506, -0.08435254544019699, -0.10592874139547348, 0.01948610506951809, -0.10643617063760757, 0.00783884059637785, -0.0925859585404396, -0.1359592080116272, 0.015173841267824173, 0.03728719428181648, -0.0568535141646862, -0.025286292657256126, -0.022119175642728806, -0.0697525143623352, -0.005578760989010334, -0.010535497218370438, 0.02044040709733963, -0.03840665891766548, 0.06527778506278992, 0.06332897394895554, 0.08601044863462448, -0.020152226090431213, 0.047389812767505646, -0.1085602194070816, 0.0373152494430542, -0.21635624766349792, 0.04892248287796974, -0.06737223267555237, 0.03794930875301361, -0.1137811541557312, -0.12576405704021454, 0.0797969251871109, -0.031367648392915726, 0.09887264668941498, 0.1270117312669754, -0.12583310902118683, -0.05015905946493149, 0.17255181074142456, -0.06802739202976227, -0.11637736111879349, 0.12964847683906555, -0.05627790465950966, 0.003965714015066624, 0.07103534042835236, 0.16545885801315308, -0.005609194282442331, -0.11517724394798279, -0.02162322960793972, -0.006423527840524912, 0.0366908498108387, -0.04016478732228279, 0.0751880630850792, 0.03082280047237873, 0.020194370299577713, 0.016043899580836296, 0.00790761224925518, -0.014975914731621742, -0.12105993181467056, -0.07519987225532532, -0.05899721011519432, -0.08693430572748184, 0.09420091658830643, 0.025590496137738228, 0.06821101903915405, -0.10115712881088257, -0.11017322540283203, 0.06673257797956467, 0.07484972476959229, -0.048485662788152695, 0.026150740683078766, -0.14186879992485046, 0.10958124697208405, -0.10774954408407211, -0.028115125373005867, -0.20688824355602264, -0.02662324346601963, 0.0437595471739769, -0.027186201885342598, 0.038898568600416183, 0.04177573323249817, 0.0645197331905365, 0.074742890894413, -0.027509376406669617, -0.016049638390541077, -0.08003289997577667, -0.014475167728960514, -0.13249322772026062, -0.13786225020885468, -0.07239489257335663, -0.042354099452495575, 0.05922437831759453, -0.15698601305484772, 0.028459375724196434, 0.008894109167158604, 0.11858619749546051, 0.02050153538584709, -0.02416941337287426, -0.013626767322421074, 0.05878585949540138, -0.014391351491212845, -0.07965986430644989, 0.05577607825398445, -0.002694790717214346, -0.09332908689975739, -0.07883575558662415, -0.14321045577526093, 0.12481905519962311, 0.11376611143350601, -0.005718866363167763, -0.06382250040769577, 0.02575351856648922, -0.06617624312639236, -0.020317016169428825, -0.053233999758958817, 0.031153518706560135, 0.2521704137325287, 0.018211130052804947, 0.1628732830286026, -0.08274044096469879, -0.05728166550397873, 0.042511869221925735, -0.032560382038354874, -0.0017709394451230764, 0.07629801332950592, 0.01628551445901394, -0.1067366823554039, 0.1037212610244751, 0.12688913941383362, -0.06266959011554718, 0.08042457699775696, -0.045395586639642715, -0.073849156498909, -0.03869514539837837, -0.03335799276828766, -0.0017390343127772212, 0.12666237354278564, -0.12764205038547516, -0.03108774684369564, 0.04948242008686066, 0.005968603305518627, 0.003372766310349107, -0.17938511073589325, -0.011733664199709892, 0.03752707690000534, -0.0034369241911917925, -0.05708286166191101, -0.007477528881281614, 0.03641391918063164, 0.10474051535129547, 0.03499240055680275, -0.0626172199845314, 0.036433830857276917, 0.011426523327827454, -0.05797358602285385, 0.18180851638317108, -0.08950748294591904, -0.15664902329444885, -0.09111060202121735, -0.043271034955978394, -0.07627664506435394, -0.0036408458836376667, 0.05527009442448616, -0.08852719515562057, -0.039384257048368454, -0.07167893648147583, 0.0009828837355598807, -0.03443619981408119, 0.0017150311032310128, 0.11850030720233917, -0.03686198219656944, 0.08711151033639908, -0.1270599067211151, -0.02689969725906849, -0.03285830467939377, -0.048129547387361526, 0.036340489983558655, 0.034496452659368515, 0.11197507381439209, 0.13003790378570557, -0.07666386663913727, 0.042634353041648865, -0.03312697634100914, 0.2795906960964203, -0.06026371568441391, -0.04335325211286545, 0.13791102170944214, -0.0054755546152591705, 0.062458012253046036, 0.08013845980167389, 0.0616825707256794, -0.10237189382314682, -0.0072628711350262165, 0.013989822939038277, -0.0382625013589859, -0.23987379670143127, -0.033489931374788284, -0.015918312594294548, -0.006639985833317041, 0.10772014409303665, 0.014958847314119339, 0.04541800543665886, 0.06969518214464188, 0.008633977733552456, 0.10009859502315521, -0.05188269168138504, 0.11419633030891418, 0.09581723064184189, 0.03658170625567436, 0.11820629239082336, -0.03690575808286667, -0.07764045149087906, 0.03949468582868576, -0.025228941813111305, 0.24671423435211182, 0.019972041249275208, 0.10511017590761185, 0.02962247096002102, 0.13142850995063782, 0.019284699112176895, 0.09156948328018188, -0.0034878647420555353, -0.025905011221766472, -0.0163920596241951, -0.04343259334564209, -0.07136133313179016, 0.0028089324478060007, -0.046146493405103683, 0.08526881039142609, -0.13230377435684204, 0.030894042924046516, 0.0209671463817358, 0.26963263750076294, 0.023024100810289383, -0.35979846119880676, -0.10853462666273117, -0.005819806829094887, -0.03855404257774353, -0.06740221381187439, 0.029422089457511902, 0.08255112916231155, -0.11292805522680283, 0.030107591301202774, -0.08284255117177963, 0.1092241108417511, -0.03667731583118439, 0.03264566510915756, 0.015110328793525696, 0.11373081058263779, 0.003561105579137802, 0.07602924853563309, -0.23798014223575592, 0.24199508130550385, 0.032747987657785416, 0.08955811709165573, -0.09133994579315186, 0.006758410949259996, -0.013026218861341476, 0.10985938459634781, 0.11083906888961792, -0.009306567721068859, -0.031072553247213364, -0.19418492913246155, -0.08046382665634155, 0.017620963975787163, 0.09195639193058014, 0.0052960701286792755, 0.10853373259305954, -0.01750202849507332, 0.01510811410844326, 0.0262383371591568, -0.01564088463783264, -0.10605952888727188, -0.09180853515863419, 0.03204689174890518, 0.029787782579660416, -0.04356655478477478, -0.06782903522253036, -0.12277384847402573, -0.05092300847172737, 0.15151356160640717, 0.06321923434734344, -0.0499957799911499, -0.12721732258796692, 0.062168609350919724, 0.08936107158660889, -0.07089947909116745, 0.007836531847715378, -0.0007722886512055993, 0.1292036920785904, -0.006703782361000776, -0.0904189944267273, 0.09020048379898071, -0.06388893723487854, -0.14807187020778656, -0.021337514743208885, 0.10628710687160492, 0.05730028450489044, 0.07235319912433624, 0.008273154497146606, 0.022018956020474434, -0.006637920159846544, -0.07968756556510925, 0.020716747269034386, 0.050507452338933945, 0.0813063457608223, 0.0021329259034246206, -0.011589495465159416, 0.05589006096124649, -0.06545660644769669, 0.045697182416915894, 0.16524051129817963, 0.22200757265090942, -0.09570640325546265, 0.06216951087117195, 0.06045379862189293, -0.047508832067251205, -0.1589488834142685, 0.04095178097486496, 0.05290050804615021, 0.009577873162925243, 0.0169967208057642, -0.15029668807983398, 0.09496431797742844, 0.0849045068025589, -0.015448537655174732, 0.032869983464479446, -0.29193049669265747, -0.09977809339761734, 0.10314838588237762, 0.12182927131652832, 0.1504303365945816, -0.12791293859481812, -0.019035596400499344, -0.016910120844841003, -0.1646157056093216, 0.07694129645824432, -0.09055975824594498, 0.10503837466239929, -0.04627056047320366, 0.13133423030376434, 0.02349170297384262, -0.05057428032159805, 0.12084930390119553, 0.023270288482308388, 0.0982503816485405, -0.0648047998547554, -0.0250942911952734, 0.16060958802700043, -0.07130331546068192, 0.078147754073143, -0.014609701931476593, 0.041661038994789124, -0.14043784141540527, -0.01748715154826641, -0.06576478481292725, 0.05276196077466011, -0.03500998020172119, -0.047011155635118484, -0.043995894491672516, 0.020679552108049393, 0.051451511681079865, -0.02397107146680355, 0.15958860516548157, 0.057387154549360275, 0.13306260108947754, 0.11863141506910324, 0.1307215541601181, -0.10753133893013, -0.09257325530052185, 0.01533168088644743, -0.011966381222009659, 0.07925989478826523, -0.12699240446090698, 0.02762414701282978, 0.14889110624790192, 0.0322478748857975, 0.11482315510511398, 0.06731373071670532, -0.05900866165757179, 0.017091549932956696, 0.007129888515919447, -0.15809588134288788, -0.15292666852474213, -0.010427897796034813, 0.04909772425889969, -0.11588648706674576, 0.09217339009046555, 0.09332586824893951, -0.09165076166391373, -0.0448722168803215, -0.014850416220724583, -0.021333232522010803, -0.022846154868602753, 0.1497497856616974, 0.04241945594549179, 0.07953337579965591, -0.1110035628080368, 0.09372541308403015, 0.06270776689052582, -0.09336283057928085, 0.05338259041309357, 0.05481056496500969, -0.08943063020706177, -0.02622036449611187, 0.06627224385738373, 0.1979905664920807, -0.06779385358095169, -0.020826688036322594, -0.12999236583709717, -0.10279809683561325, 0.06506651639938354, 0.1678832471370697, 0.0815398171544075, -0.0344279408454895, -0.04565522447228432, 0.019948482513427734, -0.16347035765647888, 0.09106409549713135, 0.029759015887975693, 0.06412161141633987, -0.14096102118492126, 0.11176689714193344, 0.004487685393542051, 0.08921205252408981, -0.03614899516105652, -0.0018637367757037282, -0.10008121281862259, 0.013389963656663895, -0.17824353277683258, -0.020139873027801514, -0.05230793356895447, 0.0034348415210843086, -0.030713491141796112, -0.04713539406657219, -0.045016903430223465, 0.042796678841114044, -0.09570135176181793, -0.012467848137021065, 0.043784741312265396, 0.016055619344115257, -0.1198597401380539, -0.013875016942620277, -0.0004900190397165716, -0.07403979450464249, 0.07833031564950943, 0.06392452865839005, -0.015717487782239914, 0.009520200081169605, -0.022011198103427887, -0.023673774674534798, 0.027069250121712685, 0.008351688273251057, 0.11849487572908401, -0.08551577478647232, -0.012042044661939144, -0.011597058735787868, 0.03653942048549652, 0.024433765560388565, 0.05904074385762215, -0.10986805707216263, -0.049643803387880325, -0.0013666516169905663, -0.05469409376382828, -0.0745447650551796, 0.05613662675023079, 0.11165639758110046, 0.02001248300075531, 0.1479072868824005, -0.058580704033374786, 0.07701674848794937, -0.17248602211475372, -0.03102525696158409, 0.0048803179524838924, -0.07606273889541626, -0.059173792600631714, -0.04954097792506218, 0.06916730105876923, -0.07306516170501709, 0.13929158449172974, 0.0034805014729499817, 0.09999403357505798, 0.03725799545645714, -0.048541903495788574, -0.03479350358247757, -0.007995286025106907, 0.16871896386146545, 0.03213769942522049, -0.03648896515369415, 0.10913404822349548, 0.02725612185895443, 0.08388342708349228, 0.09160444140434265, 0.18741706013679504, 0.12996332347393036, 0.004076845478266478, 0.08328139781951904, 0.04464764893054962, -0.026682371273636818, -0.20759953558444977, -0.008515606634318829, 0.013529200106859207, 0.10515142232179642, -0.022269872948527336, 0.17034649848937988, 0.11226102709770203, -0.11563724279403687, 0.05738737806677818, -0.059367839246988297, -0.09217754006385803, -0.11541631817817688, -0.06812376528978348, -0.07296475768089294, -0.14510218799114227, 0.024051355198025703, -0.13424591720104218, 0.015711866319179535, 0.07047698646783829, 0.005333690904080868, -0.02184680849313736, 0.1677723377943039, -0.026337239891290665, 0.03508218377828598, 0.06577856838703156, 0.0014549016486853361, -0.022472746670246124, -0.0505693182349205, -0.032191190868616104, 0.023887338116765022, -0.01737295277416706, 0.07237328588962555, -0.020202860236167908, -0.012013141065835953, 0.02271932177245617, -0.028740059584379196, -0.1057177260518074, 0.01670669950544834, 0.028776921331882477, 0.05544587969779968, 0.02597448229789734, 0.009936206974089146, -0.0033306025434285402, -0.05694393068552017, 0.21269521117210388, -0.08405552804470062, -0.036437828093767166, -0.11583281308412552, 0.2777845859527588, 0.045244112610816956, -0.026779673993587494, 0.050567805767059326, -0.07946361601352692, -0.019925279542803764, 0.2052665650844574, 0.1401960700750351, -0.022196175530552864, -0.013686125166714191, -0.031089549884200096, -0.014315123669803143, -0.05478201434016228, 0.08483371883630753, 0.11764916032552719, -0.014292771928012371, -0.08753565698862076, -0.026246551424264908, -0.06055969372391701, -0.02729509398341179, -0.08445567637681961, 0.049589645117521286, 0.04553607106208801, 0.012408057227730751, -0.060443609952926636, 0.02433011680841446, -0.0510098822414875, -0.10309431701898575, 0.03818906471133232, -0.17847585678100586, -0.14496812224388123, -0.005936067085713148, 0.054425228387117386, 0.008150679990649223, 0.09484155476093292, -0.004978321958333254, -0.004330146126449108, 0.12989869713783264, -0.004974664654582739, -0.07844742387533188, -0.07350967824459076, 0.12079787254333496, -0.14536763727664948, 0.19743786752223969, 0.00041328201768919826, 0.08251455426216125, 0.12343116104602814, 0.031705670058727264, -0.1339770257472992, 0.033381666988134384, 0.04352106526494026, -0.03378007188439369, -0.020474744960665703, 0.1637967824935913, -0.011322923004627228, 0.06498641520738602, 0.025399042293429375, -0.1681104302406311, -0.043459102511405945, -0.05463948845863342, 0.012391587719321251, -0.05200988054275513, -0.013826687820255756, -0.04635145142674446, 0.13468648493289948, 0.20758169889450073, -0.04421783983707428, -0.014929726719856262, -0.06413748115301132, 0.028712784871459007, 0.07552910596132278, -0.000038164842408150434, -0.05307115614414215, -0.22566774487495422, 0.012815778143703938, 0.045179516077041626, -0.004108451306819916, -0.2405480444431305, -0.08711020648479462, 0.022963043302297592, -0.06266912817955017, -0.07345325499773026, 0.05651746317744255, 0.06270690262317657, 0.05008644610643387, -0.05213596671819687, -0.0893433466553688, -0.06974009424448013, 0.15507864952087402, -0.15135104954242706, -0.04463657736778259 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ai-club-inductions-21-nlp-ELECTRA-base-squad This model is the deepset/electra-base-squad2 pre-trained model trained on data from AI Inductions 21 NLP competition (https://www.kaggle.com/c/ai-inductions-21-nlp) for extractive QA. ## Model description More information needed ## Intended uses & limitations AI Inductions 21 NLP competition ## Training and evaluation data AI Inductions 21 NLP competition data ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - max_length = 512 - doc_stride = 384 - learning_rate: 2e-05 - weight_decay=0.01 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cpu - Datasets 1.14.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "ai-club-inductions-21-nlp-ELECTRA-base-squad", "results": []}]}
question-answering
AyushPJ/ai-club-inductions-21-nlp-ELECTRA-base-squad
[ "transformers", "pytorch", "electra", "question-answering", "generated_from_trainer", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #electra #question-answering #generated_from_trainer #endpoints_compatible #region-us
# ai-club-inductions-21-nlp-ELECTRA-base-squad This model is the deepset/electra-base-squad2 pre-trained model trained on data from AI Inductions 21 NLP competition (URL for extractive QA. ## Model description More information needed ## Intended uses & limitations AI Inductions 21 NLP competition ## Training and evaluation data AI Inductions 21 NLP competition data ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - max_length = 512 - doc_stride = 384 - learning_rate: 2e-05 - weight_decay=0.01 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cpu - Datasets 1.14.0 - Tokenizers 0.10.3
[ "# ai-club-inductions-21-nlp-ELECTRA-base-squad\n\nThis model is the deepset/electra-base-squad2 pre-trained model trained on data from AI Inductions 21 NLP competition (URL for extractive QA.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nAI Inductions 21 NLP competition", "## Training and evaluation data\n\nAI Inductions 21 NLP competition data", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- max_length = 512\n- doc_stride = 384\n- learning_rate: 2e-05\n- weight_decay=0.01\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #electra #question-answering #generated_from_trainer #endpoints_compatible #region-us \n", "# ai-club-inductions-21-nlp-ELECTRA-base-squad\n\nThis model is the deepset/electra-base-squad2 pre-trained model trained on data from AI Inductions 21 NLP competition (URL for extractive QA.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nAI Inductions 21 NLP competition", "## Training and evaluation data\n\nAI Inductions 21 NLP competition data", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- max_length = 512\n- doc_stride = 384\n- learning_rate: 2e-05\n- weight_decay=0.01\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ 37, 62, 6, 17, 14, 3, 113, 33 ]
[ "passage: TAGS\n#transformers #pytorch #electra #question-answering #generated_from_trainer #endpoints_compatible #region-us \n# ai-club-inductions-21-nlp-ELECTRA-base-squad\n\nThis model is the deepset/electra-base-squad2 pre-trained model trained on data from AI Inductions 21 NLP competition (URL for extractive QA.## Model description\n\nMore information needed## Intended uses & limitations\n\nAI Inductions 21 NLP competition## Training and evaluation data\n\nAI Inductions 21 NLP competition data## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- max_length = 512\n- doc_stride = 384\n- learning_rate: 2e-05\n- weight_decay=0.01\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ -0.07828664779663086, 0.16111969947814941, -0.0010459370678290725, 0.10766474157571793, 0.08321454375982285, 0.021705714985728264, 0.09049910306930542, 0.13325832784175873, -0.028388014063239098, 0.09588810801506042, 0.08782260119915009, 0.07184495776891708, 0.0711352601647377, 0.14545084536075592, -0.030445026233792305, -0.1727064847946167, 0.038112811744213104, 0.001346465083770454, -0.008125794120132923, 0.10139179974794388, 0.09967952221632004, -0.11909300088882446, 0.06717801094055176, 0.0361185185611248, -0.10184093564748764, 0.017643295228481293, -0.04352452605962753, -0.04325968027114868, 0.07786238938570023, -0.006205825600773096, 0.08134713768959045, -0.021051114425063133, 0.10543321073055267, -0.1809326857328415, -0.006077698897570372, 0.017384817823767662, 0.036612577736377716, 0.09505510330200195, 0.006537150591611862, 0.04196939617395401, 0.04854234680533409, -0.06240415945649147, 0.09786107391119003, 0.0011371852597221732, -0.11524844169616699, -0.13913972675800323, -0.08562272787094116, 0.045199859887361526, 0.13191884756088257, 0.11607638746500015, -0.04029806703329086, 0.19512039422988892, -0.10215122252702713, 0.039324045181274414, 0.10137052088975906, -0.3147127330303192, -0.05965675413608551, -0.03115111030638218, 0.03214731812477112, 0.03846948593854904, -0.08318352699279785, -0.041960738599300385, 0.04319610446691513, 0.04120314121246338, 0.06757257133722305, -0.009282776154577732, -0.08755487948656082, 0.008903318084776402, -0.1298985630273819, -0.06035192683339119, 0.16261976957321167, 0.07135593891143799, -0.04013660177588463, -0.13827289640903473, -0.0273162629455328, -0.18208648264408112, -0.009117595851421356, -0.044259488582611084, 0.024031244218349457, -0.03286835551261902, -0.007523463573306799, 0.055601198226213455, -0.06642644107341766, -0.04780860245227814, -0.007776149082928896, 0.12310169637203217, 0.056006625294685364, 0.02017888054251671, 0.02780694141983986, 0.10102208703756332, 0.03093479946255684, -0.13972745835781097, -0.019865458831191063, 0.015598926693201065, -0.11763136833906174, -0.0606938898563385, -0.03618278726935387, -0.008618704974651337, -0.016153214499354362, 0.15391701459884644, -0.022480260580778122, 0.029558349400758743, 0.06922613084316254, -0.014085068367421627, 0.009957523085176945, 0.12513098120689392, -0.0669526532292366, -0.03699639067053795, -0.012674137949943542, 0.11971015483140945, 0.020242424681782722, -0.0005051657790318131, -0.06999070942401886, 0.0031420071609318256, 0.07638442516326904, 0.022641807794570923, -0.04801560193300247, 0.024145759642124176, -0.13340285420417786, -0.05343589186668396, 0.08587361127138138, -0.09138589352369308, 0.0333612784743309, -0.0039174845442175865, -0.12210661172866821, -0.10787228494882584, -0.01584288664162159, 0.04439166560769081, -0.026751117780804634, 0.04166657105088234, -0.10564113408327103, 0.0026185268070548773, -0.07498694956302643, -0.08203299343585968, 0.010216505266726017, -0.029566679149866104, -0.0069947317242622375, -0.055124133825302124, -0.16670791804790497, -0.019838333129882812, 0.03251190856099129, -0.07493174076080322, -0.09815283119678497, -0.05060224235057831, -0.00479024276137352, -0.016519423574209213, 0.0040704975835978985, 0.05551791936159134, -0.057888228446245193, 0.05723380297422409, 0.02373788319528103, 0.06421244144439697, 0.027935700491070747, 0.038034588098526, -0.13076110184192657, 0.02706422097980976, -0.10905636847019196, 0.08490892499685287, -0.061256520450115204, 0.04174761474132538, -0.09801752865314484, -0.09202456474304199, 0.007722212467342615, -0.028328265994787216, 0.08534575998783112, 0.13146698474884033, -0.15195763111114502, -0.019466180354356766, 0.10065849125385284, -0.04096655920147896, -0.1316470056772232, 0.1311134546995163, -0.06654217094182968, -0.026182997971773148, 0.05863898992538452, 0.1778993308544159, 0.08200975507497787, -0.06979340314865112, -0.07316610962152481, -0.07072626054286957, 0.06320800632238388, 0.016402408480644226, 0.10818929225206375, 0.009603003039956093, 0.016221944242715836, -0.014860530383884907, -0.030216945335268974, 0.0007418899913318455, -0.07617529481649399, -0.10675659030675888, -0.002143767662346363, -0.08635660260915756, 0.03951369225978851, 0.05719553306698799, 0.053764164447784424, -0.08466005325317383, -0.10958992689847946, -0.013225666247308254, 0.12035264819860458, -0.04140520095825195, -0.01786249317228794, -0.1348806768655777, 0.06964273005723953, -0.0998397022485733, -0.040892940014600754, -0.15682530403137207, -0.05065445601940155, 0.08318571746349335, -0.0523880198597908, 0.03889908269047737, 0.06997650861740112, 0.04575575515627861, 0.05566906929016113, -0.0053247190080583096, -0.03580155223608017, -0.08156195282936096, -0.020941415801644325, -0.11831390857696533, -0.16177313029766083, -0.053418487310409546, -0.032656505703926086, 0.2201792150735855, -0.2164061963558197, -0.00830328743904829, -0.0333356074988842, 0.1236785426735878, 0.022648995742201805, -0.07797979563474655, 0.017385993152856827, 0.043866273015737534, -0.020844880491495132, -0.05381287634372711, 0.02242904342710972, -0.017371026799082756, -0.09632445126771927, -0.07205624133348465, -0.10627885162830353, 0.03144772723317146, 0.05973140895366669, 0.04125817120075226, -0.09097807109355927, 0.008765249513089657, -0.0544135645031929, -0.032224949449300766, -0.07721859961748123, 0.014258352108299732, 0.2471902072429657, 0.0022893710993230343, 0.1097501888871193, -0.0748135969042778, -0.07806213945150375, -0.01314662117511034, 0.01351043488830328, 0.02863634005188942, 0.1064695417881012, 0.07460244745016098, -0.15395204722881317, 0.09432286769151688, 0.040912698954343796, -0.03223489969968796, 0.11290009319782257, -0.029355259612202644, -0.08473803848028183, -0.05257413908839226, 0.024189885705709457, 0.0030158942099660635, 0.11411324888467789, -0.08753269165754318, 0.05413106456398964, 0.05233023688197136, 0.00023556288215331733, 0.0022425532806664705, -0.1702791452407837, -0.03128587082028389, 0.04423021152615547, -0.045564256608486176, -0.027018576860427856, -0.02315262146294117, 0.028206799179315567, 0.10447561740875244, 0.03294995799660683, -0.04891294240951538, 0.012315571308135986, -0.026481078937649727, -0.08907806873321533, 0.1694919466972351, -0.0791931077837944, -0.15428264439105988, -0.10213568806648254, 0.0661613792181015, -0.058027856051921844, 0.011542567983269691, 0.009185602888464928, -0.050960808992385864, -0.061215467751026154, -0.1198970228433609, -0.06891267746686935, 0.03357727453112602, -0.01373791228979826, 0.10664194822311401, -0.035552263259887695, 0.07073154300451279, -0.10974118113517761, 0.007956494577229023, -0.032240062952041626, -0.012143990024924278, -0.004017316270619631, 0.047913920134305954, 0.10200458019971848, 0.1235070526599884, -0.01018436998128891, 0.013717697001993656, -0.006470180116593838, 0.3058713376522064, -0.09875302016735077, -0.043099869042634964, 0.08173435181379318, -0.015568441711366177, 0.015500780194997787, 0.15939146280288696, 0.03493519499897957, -0.10656620562076569, 0.03469246253371239, 0.019252756610512733, 0.0010681676212698221, -0.21156130731105804, -0.06775442510843277, -0.04342305660247803, -0.0755571722984314, 0.11961537599563599, 0.028099024668335915, 0.013541806489229202, 0.06559105217456818, -0.0027065568137913942, 0.052722539752721786, -0.038318634033203125, 0.059928420931100845, 0.09344250708818436, 0.07047911733388901, 0.10897732526063919, -0.043949421495199203, -0.09008466452360153, 0.06321274489164352, -0.007229565642774105, 0.2490629404783249, -0.009093457832932472, 0.17237898707389832, 0.025951925665140152, 0.16799212992191315, 0.010111362673342228, 0.04688269644975662, 0.015603248961269855, -0.020911963656544685, -0.008103633299469948, -0.01961561292409897, -0.029393130913376808, 0.058459874242544174, 0.07700512558221817, 0.05738954618573189, -0.05458783358335495, 0.007442417088896036, 0.024942241609096527, 0.2713741064071655, 0.07298416644334793, -0.3322100341320038, -0.04408792406320572, 0.026670482009649277, -0.07426223158836365, -0.03947769105434418, 0.0015404138248413801, 0.10869163274765015, -0.11128658056259155, 0.059195004403591156, -0.04351707920432091, 0.07261853665113449, -0.07062246650457382, -0.008071549236774445, 0.06672556698322296, 0.06904744356870651, 0.004706914536654949, 0.09283404052257538, -0.18462026119232178, 0.20283643901348114, 0.010014399886131287, 0.10122901946306229, -0.09338515251874924, 0.026556821539998055, -0.025075197219848633, -0.004851037170737982, 0.11121566593647003, -0.007598632946610451, -0.024417178705334663, -0.17505621910095215, -0.09317201375961304, 0.04629136994481087, 0.024498343467712402, -0.07076811045408249, 0.11416691541671753, -0.012158440425992012, 0.010201833210885525, 0.011950920335948467, -0.014299903996288776, -0.07345648109912872, -0.11929840594530106, 0.03401721641421318, 0.053469713777303696, -0.008870971389114857, -0.050602443516254425, -0.09004291892051697, -0.08574764430522919, 0.16103726625442505, -0.017099745571613312, -0.03618057444691658, -0.11460475623607635, 0.0802016407251358, 0.11289456486701965, -0.05665786191821098, -0.004092673305422068, 0.014689251780509949, 0.0857241228222847, 0.024513330310583115, -0.01529653649777174, 0.08209116011857986, -0.04938582330942154, -0.1779305636882782, -0.03361804038286209, 0.08953599631786346, 0.05074604973196983, 0.08997205644845963, 0.014790698885917664, 0.03912843391299248, -0.02099502459168434, -0.08694819360971451, 0.012622678652405739, 0.0060753352008759975, 0.06910542398691177, 0.05518559366464615, -0.011706891469657421, -0.012270724400877953, -0.0727069154381752, -0.004230131860822439, 0.1261557936668396, 0.28848373889923096, -0.08842972666025162, 0.03909200429916382, 0.01796175353229046, -0.045647166669368744, -0.10211879014968872, 0.01881369948387146, 0.08755501359701157, 0.03853367641568184, 0.10671867430210114, -0.13366901874542236, 0.08805574476718903, 0.11207496374845505, -0.031001422554254532, 0.0512864850461483, -0.2996329963207245, -0.12636956572532654, 0.04879241809248924, 0.10413163155317307, 0.05384724214673042, -0.07033704966306686, -0.02462012693285942, 0.016605263575911522, -0.12515245378017426, 0.09137769043445587, 0.05354113131761551, 0.1183067113161087, -0.009108087047934532, 0.05228744074702263, 0.03324681147933006, -0.04698589816689491, 0.16110052168369293, 0.02483098953962326, 0.09446042776107788, -0.03719749674201012, 0.004328327719122171, 0.01718614436686039, -0.031900011003017426, 0.08684055507183075, -0.04239319637417793, 0.027449533343315125, -0.2511254549026489, -0.022802988067269325, -0.0501650795340538, 0.01603970304131508, -0.02666962705552578, -0.039288341999053955, -0.07587431371212006, 0.062259089201688766, 0.07258874922990799, 0.0068654464557766914, 0.08320356905460358, 0.04146747663617134, 0.10514991730451584, 0.039668720215559006, 0.1223808079957962, 0.05498195439577103, -0.0375373475253582, -0.017279809340834618, 0.02012098766863346, 0.052270881831645966, -0.05947801098227501, 0.030443839728832245, 0.16307230293750763, 0.010554579086601734, 0.15445436537265778, 0.017773689702153206, -0.06588524580001831, 0.022854948416352272, 0.054872091859579086, -0.1467745155096054, -0.12258700281381607, -0.00263961567543447, -0.044109925627708435, -0.12853562831878662, -0.04491683840751648, 0.08227822184562683, -0.0638175681233406, -0.029855042695999146, 0.014471162110567093, 0.00734751159325242, -0.04465775191783905, 0.1961858570575714, -0.006310300435870886, 0.07600850611925125, -0.10006671398878098, 0.08963264524936676, 0.10364781320095062, -0.045317258685827255, 0.05993419513106346, 0.054539959877729416, -0.05856747180223465, -0.01933032087981701, 0.00597462197765708, 0.18843010067939758, 0.017382478341460228, -0.029924005270004272, -0.14195244014263153, -0.09072992950677872, 0.07470307499170303, 0.04081740975379944, 0.06912226229906082, -0.00626353221014142, -0.009411501698195934, -0.0023542961571365595, -0.10906548798084259, 0.10947181284427643, 0.06737258285284042, 0.04123248532414436, -0.10523416846990585, 0.05922564119100571, -0.010851068422198296, 0.01259684469550848, -0.01784752868115902, 0.013513599522411823, -0.07003813982009888, -0.006956955883651972, -0.06706104427576065, -0.03488631919026375, -0.065091073513031, 0.0009334274800494313, -0.017911987379193306, -0.06416281312704086, -0.06524166464805603, 0.0032572797499597073, -0.09184426069259644, -0.026219911873340607, 0.014417956583201885, 0.05062543600797653, -0.1372489333152771, -0.042230792343616486, 0.029400045052170753, -0.07049961388111115, 0.09254399687051773, 0.008600565604865551, 0.019878949970006943, -0.03742580488324165, -0.030296361073851585, 0.011816266924142838, 0.02968583069741726, 0.04155627265572548, 0.06177058815956116, -0.14911353588104248, -0.00022846848878543824, -0.004092917777597904, 0.01574709452688694, 0.013587797991931438, 0.03102194145321846, -0.11542628705501556, -0.056321658194065094, -0.021666035056114197, -0.05975151062011719, -0.06708220392465591, 0.025204189121723175, 0.09337738901376724, 0.025298980996012688, 0.21011270582675934, -0.03923863545060158, 0.05753151327371597, -0.2017960101366043, -0.03502370044589043, 0.010821505449712276, -0.016043810173869133, -0.10486920177936554, -0.011899753473699093, 0.07522918283939362, -0.0484069399535656, 0.13962586224079132, -0.008931579999625683, 0.006916339509189129, 0.048652682453393936, 0.011725110001862049, 0.013788406737148762, 0.010214577428996563, 0.13332045078277588, 0.05293242260813713, -0.03676239401102066, 0.08196379244327545, -0.025970906019210815, 0.05356651172041893, 0.03197832778096199, 0.10395175963640213, 0.16321200132369995, 0.037662286311388016, 0.048187255859375, 0.04252679646015167, -0.0602434016764164, -0.17255771160125732, 0.10396264493465424, -0.06465911120176315, 0.09198832511901855, 0.0030868016183376312, 0.17947229743003845, 0.04490695893764496, -0.1886453926563263, 0.039151813834905624, -0.04542131349444389, -0.09358904510736465, -0.0746549591422081, -0.1212475597858429, -0.07798699289560318, -0.11408454924821854, 0.02930188551545143, -0.13101628422737122, 0.05154269561171532, 0.13592559099197388, 0.022067155689001083, -0.027614831924438477, 0.14281678199768066, -0.05031515657901764, 0.055720504373311996, 0.04984787106513977, -0.0041047814302146435, 0.0045950766652822495, -0.04320113733410835, -0.05102725699543953, 0.01423908956348896, 0.009961850009858608, 0.09810225665569305, -0.0442199669778347, -0.04946908354759216, 0.022564653307199478, 0.021402910351753235, -0.10234205424785614, 0.021113406866788864, -0.013760292902588844, 0.05195698142051697, 0.08912196010351181, 0.0519532784819603, 0.03846540302038193, -0.033557742834091187, 0.1930936574935913, -0.05807313695549965, -0.021062251180410385, -0.1927001178264618, 0.13625797629356384, 0.007175631355494261, -0.029776556417346, 0.08321909606456757, -0.08440330624580383, 0.0048376889899373055, 0.17333540320396423, 0.08623937517404556, -0.10575084388256073, -0.022159237414598465, -0.02797587215900421, -0.02066054753959179, -0.04703386500477791, 0.07354042679071426, 0.05558636412024498, 0.014759412966668606, -0.08262095600366592, -0.018202785402536392, -0.07784934341907501, -0.02391253598034382, -0.04770568758249283, 0.01784345880150795, 0.048881761729717255, 0.009138116613030434, -0.0619812086224556, 0.0378861241042614, -0.0448634959757328, -0.10279268771409988, 0.04237690567970276, -0.17447307705879211, -0.1899169385433197, -0.029941895976662636, 0.057817742228507996, 0.0003007907362189144, 0.05124005675315857, -0.012113847769796848, -0.005886750761419535, 0.06837954372167587, -0.00951527152210474, -0.04591963440179825, -0.1006602793931961, 0.11393197625875473, -0.07998834550380707, 0.19484388828277588, -0.008661163970828056, 0.10827573388814926, 0.09665703773498535, 0.0191242303699255, -0.11753562092781067, 0.06125596910715103, 0.06764943152666092, -0.04010528698563576, 0.018533945083618164, 0.14698421955108643, -0.0012191623682156205, 0.1462104618549347, 0.07743214070796967, -0.10015681385993958, -0.000899696140550077, -0.11011762171983719, 0.017851047217845917, -0.0988541692495346, -0.006515394430607557, -0.05078510567545891, 0.1748618483543396, 0.19604316353797913, -0.046284548938274384, 0.00881904736161232, -0.04007897526025772, 0.03617127984762192, 0.030284220352768898, 0.08608279377222061, -0.05090200528502464, -0.2482939511537552, 0.02026115544140339, 0.03252594918012619, 0.014610044658184052, -0.19939956068992615, -0.10065904259681702, -0.028184127062559128, -0.08268570154905319, -0.030490422621369362, 0.10281549394130707, 0.080491803586483, 0.038880739361047745, -0.04351671040058136, -0.10418090224266052, -0.03757930546998978, 0.11428240686655045, -0.12472722679376602, -0.06509038060903549 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ai-club-inductions-21-nlp-XLNet This model was trained from scratch on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cpu - Datasets 1.14.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "ai-club-inductions-21-nlp-XLNet", "results": []}]}
question-answering
AyushPJ/ai-club-inductions-21-nlp-XLNet
[ "transformers", "pytorch", "xlnet", "question-answering", "generated_from_trainer", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #xlnet #question-answering #generated_from_trainer #endpoints_compatible #region-us
# ai-club-inductions-21-nlp-XLNet This model was trained from scratch on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cpu - Datasets 1.14.0 - Tokenizers 0.10.3
[ "# ai-club-inductions-21-nlp-XLNet\n\nThis model was trained from scratch on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #xlnet #question-answering #generated_from_trainer #endpoints_compatible #region-us \n", "# ai-club-inductions-21-nlp-XLNet\n\nThis model was trained from scratch on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ 37, 30, 6, 12, 8, 3, 90, 33 ]
[ "passage: TAGS\n#transformers #pytorch #xlnet #question-answering #generated_from_trainer #endpoints_compatible #region-us \n# ai-club-inductions-21-nlp-XLNet\n\nThis model was trained from scratch on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ -0.0951448604464531, 0.09074791520833969, -0.002098863711580634, 0.08363977074623108, 0.11835363507270813, -0.02251383848488331, 0.1393555849790573, 0.11690462380647659, -0.09293117374181747, 0.05294239893555641, 0.11471915990114212, 0.09604848176240921, 0.03138379752635956, 0.16472971439361572, -0.03951895982027054, -0.21122680604457855, 0.04480869695544243, 0.07089938968420029, -0.030099332332611084, 0.12442860007286072, 0.09272460639476776, -0.13962173461914062, 0.08521337807178497, 0.04219203442335129, -0.19161823391914368, 0.011861992999911308, -0.02144426852464676, -0.08632665872573853, 0.09959118813276291, -0.010835899040102959, 0.11098185926675797, 0.0062429532408714294, 0.07492710649967194, -0.15439319610595703, 0.007074284367263317, 0.05857913941144943, 0.023649245500564575, 0.10235054045915604, 0.022926144301891327, 0.0049284351989626884, 0.09076973795890808, -0.11330176144838333, 0.07676693052053452, 0.020095564424991608, -0.08981449902057648, -0.19815140962600708, -0.08529208600521088, 0.10303360223770142, 0.06035907194018364, 0.11082153022289276, 0.003115589264780283, 0.26937368512153625, -0.06659345328807831, 0.09762468934059143, 0.2586670517921448, -0.3347229063510895, -0.06967651844024658, 0.040133073925971985, 0.08397332578897476, 0.07354331016540527, -0.08106983453035355, -0.019027069211006165, 0.06010337918996811, 0.06198839098215103, 0.12651219964027405, -0.025340018793940544, -0.03918064385652542, -0.012531316839158535, -0.15341022610664368, -0.041442982852458954, 0.17504924535751343, 0.022889193147420883, -0.05331522598862648, -0.07618799060583115, -0.08721131831407547, -0.1294068694114685, -0.027757547795772552, -0.07233186811208725, 0.031144844368100166, -0.059087734669446945, -0.0754927322268486, -0.04249636083841324, -0.06048304960131645, -0.03427743539214134, -0.05471654608845711, 0.1726660430431366, 0.05432276055216789, 0.05522703751921654, -0.08667236566543579, 0.07349397987127304, -0.012689967639744282, -0.1118604764342308, -0.01864299178123474, -0.018320206552743912, -0.0029447718989104033, -0.04629625380039215, -0.0221366286277771, -0.0683395266532898, -0.006176301743835211, 0.14844463765621185, -0.10201766341924667, 0.03898348659276962, 0.0049095721915364265, 0.029580900445580482, -0.01525521744042635, 0.1637873649597168, -0.057902202010154724, -0.002299571642652154, 0.023053418844938278, 0.06339816004037857, 0.0319022573530674, -0.023517541587352753, -0.10978484898805618, -0.0016669165343046188, 0.11984147131443024, 0.04286808893084526, -0.03612452745437622, 0.0569041445851326, -0.006580736953765154, -0.05715110898017883, 0.048577889800071716, -0.11319947987794876, 0.0451325997710228, -0.01975162699818611, -0.09661995619535446, 0.013665316626429558, -0.005270291119813919, -0.002588051138445735, -0.06862559914588928, 0.01106928288936615, -0.10497647523880005, 0.03653766214847565, -0.07878631353378296, -0.10326840728521347, 0.018858259543776512, -0.08559175580739975, -0.012431195937097073, -0.09803453832864761, -0.1446201205253601, 0.00244147889316082, 0.03909975290298462, -0.040073003619909286, -0.03319510817527771, -0.032803039997816086, -0.08881822228431702, -0.005340240430086851, -0.009686989709734917, 0.020543282851576805, -0.0370902381837368, 0.060739461332559586, 0.050694625824689865, 0.07510685920715332, -0.024458153173327446, 0.03292074054479599, -0.1217903271317482, 0.026035433635115623, -0.19230656325817108, 0.05726749822497368, -0.07002630084753036, 0.037605781108140945, -0.10867544263601303, -0.1334177553653717, 0.05266699194908142, -0.04177326709032059, 0.09336286038160324, 0.12017686665058136, -0.13185234367847443, -0.04139866679906845, 0.1666897088289261, -0.0863039642572403, -0.11655247211456299, 0.12504462897777557, -0.047841306775808334, 0.014469674788415432, 0.07378216087818146, 0.18362204730510712, 0.026963628828525543, -0.11190734058618546, -0.03117818757891655, -0.0033645418006926775, 0.007620680145919323, -0.06295029073953629, 0.04659876227378845, 0.0515967532992363, 0.057823482900857925, 0.008889746852219105, -0.022160526365041733, -0.011052669957280159, -0.10889764875173569, -0.09137752652168274, -0.06469881534576416, -0.08995179831981659, 0.0789366289973259, 0.03340741991996765, 0.06932045519351959, -0.09044480323791504, -0.09924685209989548, 0.056798055768013, 0.0909702405333519, -0.06110266223549843, 0.03993402421474457, -0.1439041942358017, 0.09924328327178955, -0.09159667044878006, -0.02136850170791149, -0.2048916071653366, -0.060045212507247925, 0.0432155579328537, -0.019417978823184967, 0.043717652559280396, 0.03775575011968613, 0.05475855618715286, 0.056846339255571365, -0.0191013365983963, -0.014590395614504814, -0.057226963341236115, -0.019728481769561768, -0.13047580420970917, -0.1536893993616104, -0.06287362426519394, -0.03777264058589935, 0.05565531551837921, -0.18834424018859863, 0.03160104528069496, 0.014528556726872921, 0.12259313464164734, 0.012481684796512127, -0.020645618438720703, -0.010273260995745659, 0.059715498238801956, -0.02864655666053295, -0.08039242774248123, 0.04762585088610649, 0.006863806862384081, -0.10637768357992172, -0.06669271737337112, -0.1258196234703064, 0.1399381160736084, 0.11330655962228775, -0.013903031125664711, -0.04909408837556839, 0.038418009877204895, -0.0677105262875557, -0.03235086798667908, -0.03661131113767624, 0.051600418984889984, 0.2282657027244568, 0.006034386344254017, 0.1434732973575592, -0.09336724132299423, -0.06344670057296753, 0.040264517068862915, -0.03163961321115494, -0.003126739989966154, 0.06944873929023743, 0.01750289835035801, -0.15453414618968964, 0.10906506329774857, 0.09025845676660538, -0.06971120089292526, 0.09244529902935028, -0.04413226991891861, -0.05124017968773842, -0.03906607627868652, -0.005635574460029602, -0.0030926528852432966, 0.11968784779310226, -0.12720629572868347, -0.025114068761467934, 0.037191856652498245, 0.010872014798223972, 0.02569672092795372, -0.17986245453357697, -0.01273657288402319, 0.016842544078826904, -0.02831980213522911, -0.05378345027565956, -0.011420625261962414, 0.030397256836295128, 0.09893637150526047, 0.022761145606637, -0.07756607234477997, 0.04709997773170471, 0.00465058209374547, -0.06934469193220139, 0.18354953825473785, -0.09788545221090317, -0.1928900182247162, -0.07795259356498718, -0.0007592469919472933, -0.10042644292116165, -0.02312747947871685, 0.034193601459264755, -0.08797595649957657, -0.03778568655252457, -0.07973691076040268, -0.04485839977860451, -0.04474765807390213, 0.008302840404212475, 0.09909068793058395, -0.011915338225662708, 0.08800587058067322, -0.13644789159297943, -0.01417507417500019, -0.05067573860287666, -0.04560256004333496, 0.026386326178908348, 0.03842407464981079, 0.11413358151912689, 0.1457824856042862, -0.07087426632642746, 0.04386331886053085, -0.020833680406212807, 0.27820122241973877, -0.07630712538957596, -0.02389509603381157, 0.12728944420814514, 0.01642443984746933, 0.0696946308016777, 0.08806323260068893, 0.05826188996434212, -0.09466029703617096, -0.006531126331537962, 0.009137015789747238, -0.037455812096595764, -0.2512194514274597, -0.05007432773709297, -0.026684191077947617, -0.02752239815890789, 0.07951070368289948, 0.038097597658634186, 0.07389611005783081, 0.06156311556696892, 0.003665799740701914, 0.06613010913133621, -0.049999043345451355, 0.10887863487005234, 0.11218544095754623, 0.05348501726984978, 0.12043048441410065, -0.04298185184597969, -0.0728481188416481, 0.02924426831305027, -0.0007501008221879601, 0.24120132625102997, -0.0006424785824492574, 0.11476688832044601, 0.036503978073596954, 0.17979680001735687, 0.017010577023029327, 0.07379709929227829, -0.013567627407610416, -0.02500421740114689, 0.004798661917448044, -0.04498309642076492, -0.043562375009059906, -0.003893042914569378, -0.043274592608213425, 0.07141652703285217, -0.09615908563137054, 0.04897972568869591, 0.032652560621500015, 0.2543881833553314, 0.008235248737037182, -0.34526100754737854, -0.09752783924341202, -0.018839159980416298, -0.04933784902095795, -0.05410725995898247, 0.019805394113063812, 0.11282283067703247, -0.11798641830682755, 0.038746368139982224, -0.08267583698034286, 0.10224445164203644, -0.02976853959262371, 0.0272810235619545, 0.027006518095731735, 0.11035320907831192, 0.007491739466786385, 0.08090589195489883, -0.21415747702121735, 0.23593257367610931, 0.025521786883473396, 0.10638511925935745, -0.09197866171598434, -0.0008501399424858391, -0.02785436436533928, 0.12473542988300323, 0.09916810691356659, 0.00008998969860840589, -0.04545950889587402, -0.15708643198013306, -0.07837697118520737, 0.01982257142663002, 0.09088248014450073, 0.039310138672590256, 0.11048074811697006, -0.012364811263978481, 0.013315434567630291, 0.031865768134593964, -0.03236156702041626, -0.10078954696655273, -0.0964830219745636, 0.02501746453344822, 0.07211880385875702, -0.056055858731269836, -0.06170658394694328, -0.11165317893028259, -0.03563462942838669, 0.20699024200439453, 0.018924539908766747, -0.06390617042779922, -0.11053335666656494, 0.04505607485771179, 0.10260050743818283, -0.07275987416505814, 0.004964380990713835, 0.0002265249058837071, 0.12167751789093018, -0.003738914616405964, -0.09445077180862427, 0.07915448397397995, -0.0701526328921318, -0.14108367264270782, -0.013860231265425682, 0.11001775413751602, 0.05856481194496155, 0.058780234307050705, 0.01602642424404621, 0.0058992644771933556, -0.024703269824385643, -0.08684083074331284, -0.0031802786979824305, 0.03591691702604294, 0.07021740078926086, 0.004408650100231171, 0.010788860730826855, 0.04485958814620972, -0.07795942574739456, 0.03675561770796776, 0.16383664309978485, 0.19056305289268494, -0.09767994284629822, 0.06772878021001816, 0.0740870013833046, -0.049626462161540985, -0.15545952320098877, 0.036531321704387665, 0.052014146000146866, -0.00011181712034158409, 0.0030519282445311546, -0.14112895727157593, 0.10194050520658493, 0.06956372410058975, -0.016054680570960045, 0.05392403155565262, -0.3236553370952606, -0.09192944318056107, 0.0749882236123085, 0.09708192944526672, 0.1374080330133438, -0.12303543835878372, -0.020676618441939354, -0.005667725577950478, -0.11926459521055222, 0.10677351802587509, -0.09611273556947708, 0.10590474307537079, -0.03553497791290283, 0.14297616481781006, 0.01549583114683628, -0.05457620322704315, 0.10587628930807114, 0.028840893879532814, 0.10184913128614426, -0.06133564934134483, -0.01785941794514656, 0.16653971374034882, -0.0611884705722332, 0.08962731808423996, -0.015742862597107887, 0.044964779168367386, -0.11475742608308792, -0.011436816304922104, -0.0809803158044815, 0.05689765885472298, -0.025025581941008568, -0.04443565756082535, -0.06555159389972687, 0.02630705013871193, 0.06251589208841324, -0.012106243520975113, 0.17181485891342163, 0.06872756034135818, 0.13363324105739594, 0.14667262136936188, 0.1352284997701645, -0.11112125962972641, -0.10656499862670898, 0.019707094877958298, -0.007139324676245451, 0.08913619071245193, -0.10209643840789795, 0.02614673040807247, 0.14664264023303986, 0.028599821031093597, 0.09145836532115936, 0.06740042567253113, -0.06889811903238297, 0.00466507812961936, 0.015294782817363739, -0.15426352620124817, -0.16255781054496765, -0.0026082652620971203, 0.021735385060310364, -0.1170518696308136, 0.12089322507381439, 0.10372409969568253, -0.09869398921728134, -0.03765975683927536, -0.013760106638073921, -0.012293234467506409, -0.030529458075761795, 0.15636512637138367, 0.061019688844680786, 0.07453589886426926, -0.11231699585914612, 0.09428058564662933, 0.07201877981424332, -0.06689348071813583, 0.06102977320551872, 0.08254653215408325, -0.10041133314371109, -0.0318099781870842, 0.06066368520259857, 0.15288761258125305, -0.08151378482580185, -0.04808666929602623, -0.1292772889137268, -0.09098410606384277, 0.06138312444090843, 0.16185787320137024, 0.07655127346515656, -0.023253390565514565, -0.04642421379685402, 0.017835678532719612, -0.16538627445697784, 0.08253529667854309, 0.02480420470237732, 0.06013762205839157, -0.16834452748298645, 0.12611474096775055, 0.009809955954551697, 0.09595376253128052, -0.034344155341386795, -0.003586631501093507, -0.09412483870983124, 0.02266216091811657, -0.13643479347229004, -0.020517829805612564, -0.05786249786615372, -0.001116391853429377, -0.027165979146957397, -0.05063190311193466, -0.06327255815267563, 0.06091035157442093, -0.08349430561065674, -0.007854223251342773, 0.04356209188699722, -0.0059357150457799435, -0.13545745611190796, -0.0047689336352050304, -0.013137515634298325, -0.06666336208581924, 0.07234670221805573, 0.06628082692623138, -0.023145120590925217, 0.002834718208760023, -0.03779042884707451, -0.012461722828447819, 0.03256222978234291, 0.022300809621810913, 0.10836423933506012, -0.06756140291690826, -0.01691601052880287, 0.015937892720103264, 0.04544670134782791, 0.02888612449169159, 0.04719995707273483, -0.10221178084611893, -0.03907706215977669, -0.023347515612840652, -0.0419686920940876, -0.06803686171770096, 0.05774335935711861, 0.11240656673908234, 0.025380535051226616, 0.16550366580486298, -0.056130342185497284, 0.05863254517316818, -0.16941314935684204, -0.026345761492848396, 0.009817683137953281, -0.07626109570264816, -0.04239466413855553, -0.027060583233833313, 0.06815546005964279, -0.07338959723711014, 0.13709145784378052, 0.007545108906924725, 0.07583548873662949, 0.029172204434871674, -0.04705491289496422, -0.05335177853703499, -0.006239033304154873, 0.1541525423526764, 0.0412561409175396, -0.03378146514296532, 0.12204387784004211, 0.008965066634118557, 0.06898896396160126, 0.10364487022161484, 0.18414899706840515, 0.10940670967102051, 0.0344868041574955, 0.0945194661617279, 0.04718488082289696, -0.02630501799285412, -0.20951968431472778, -0.00031073292484506965, -0.02416541427373886, 0.10368581861257553, -0.029081109911203384, 0.15340566635131836, 0.13510756194591522, -0.11164319515228271, 0.04648074507713318, -0.042059171944856644, -0.10053691267967224, -0.1157403290271759, -0.12330523133277893, -0.07177543640136719, -0.15135237574577332, 0.01113828457891941, -0.13372394442558289, 0.013149071484804153, 0.04028153792023659, 0.010931429453194141, -0.030363233759999275, 0.15390202403068542, 0.013739338144659996, 0.022926559671759605, 0.05245014280080795, 0.0005434444756247103, -0.026609018445014954, -0.026219403371214867, -0.013151267543435097, 0.017829790711402893, -0.01782187819480896, 0.08408903330564499, -0.011272471398115158, 0.0008517072419635952, 0.04727857559919357, -0.03060889057815075, -0.09198940545320511, 0.008702855557203293, 0.02898467518389225, 0.055753160268068314, 0.03270263969898224, 0.027159176766872406, -0.015946459025144577, -0.059231117367744446, 0.21132735908031464, -0.07839067280292511, -0.06890784204006195, -0.1154637485742569, 0.2672385275363922, 0.03367678076028824, -0.030970769003033638, 0.035728342831134796, -0.07721976190805435, -0.029970962554216385, 0.22134004533290863, 0.17497317492961884, -0.03192424029111862, -0.012283165007829666, -0.027068516239523888, -0.019935542717576027, -0.06790582090616226, 0.08116544038057327, 0.12935207784175873, 0.05341110751032829, -0.07691425085067749, -0.013852539472281933, -0.07183980941772461, -0.015600121580064297, -0.1010754331946373, 0.026700763031840324, 0.03418286144733429, 0.01616833172738552, -0.06251094490289688, 0.03528885915875435, -0.0745280534029007, -0.08523515611886978, 0.03309127315878868, -0.16647334396839142, -0.1392516791820526, 0.0028007407672703266, 0.06652836501598358, -0.008485765196383, 0.08068228513002396, -0.009119992144405842, 0.0016765845939517021, 0.09528177976608276, -0.008301105350255966, -0.07856474071741104, -0.07699188590049744, 0.12367492914199829, -0.13883981108665466, 0.19563283026218414, -0.010634507052600384, 0.09754646569490433, 0.11908736824989319, 0.017942534759640694, -0.13771061599254608, 0.03055124543607235, 0.04902327060699463, -0.030982527881860733, -0.012257262133061886, 0.13977700471878052, -0.0002092529321089387, 0.07294806838035583, 0.03581221401691437, -0.13437017798423767, -0.04959410801529884, -0.04859200492501259, 0.0417858250439167, -0.06344113498926163, -0.022739173844456673, -0.04863392189145088, 0.1377948373556137, 0.18678346276283264, -0.05264328047633171, 0.00026981253176927567, -0.06958267092704773, 0.010896116495132446, 0.05352885276079178, 0.02201206237077713, -0.03948983922600746, -0.2503677010536194, 0.014588681980967522, 0.04508906602859497, 0.012614939361810684, -0.20770426094532013, -0.08487898856401443, 0.023644477128982544, -0.06689930707216263, -0.07434076070785522, 0.05364972725510597, 0.07234369963407516, 0.03546169772744179, -0.04794846102595329, -0.05118871107697487, -0.07185256481170654, 0.14071129262447357, -0.14354076981544495, -0.04633933678269386 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ai-club-inductions-21-nlp-distilBERT This model was trained from scratch on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cu110 - Datasets 1.14.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "ai-club-inductions-21-nlp-distilBERT", "results": []}]}
question-answering
AyushPJ/ai-club-inductions-21-nlp-distilBERT
[ "transformers", "pytorch", "distilbert", "question-answering", "generated_from_trainer", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #distilbert #question-answering #generated_from_trainer #endpoints_compatible #region-us
# ai-club-inductions-21-nlp-distilBERT This model was trained from scratch on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cu110 - Datasets 1.14.0 - Tokenizers 0.10.3
[ "# ai-club-inductions-21-nlp-distilBERT\n\nThis model was trained from scratch on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cu110\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #distilbert #question-answering #generated_from_trainer #endpoints_compatible #region-us \n", "# ai-club-inductions-21-nlp-distilBERT\n\nThis model was trained from scratch on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cu110\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ 38, 32, 6, 12, 8, 3, 90, 33 ]
[ "passage: TAGS\n#transformers #pytorch #distilbert #question-answering #generated_from_trainer #endpoints_compatible #region-us \n# ai-club-inductions-21-nlp-distilBERT\n\nThis model was trained from scratch on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cu110\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ -0.10988680273294449, 0.09965062886476517, -0.002183531178161502, 0.09412699192762375, 0.12695921957492828, -0.012902225367724895, 0.14519038796424866, 0.11794862151145935, -0.09022323042154312, 0.06675529479980469, 0.1288354992866516, 0.10475093126296997, 0.02821465954184532, 0.16292212903499603, -0.04583361744880676, -0.2294231653213501, 0.039014432579278946, 0.06877560168504715, 0.005542895291000605, 0.11993025988340378, 0.0917874127626419, -0.15089038014411926, 0.08168905973434448, 0.03926035761833191, -0.20922066271305084, 0.0029666621703654528, -0.017995480448007584, -0.08960951119661331, 0.0962793380022049, -0.0218871608376503, 0.11635567247867584, 0.023201467469334602, 0.09232600778341293, -0.12449152767658234, 0.0066135902889072895, 0.03697068616747856, 0.032300617545843124, 0.10592126846313477, 0.021761266514658928, -0.009578552097082138, 0.0703679546713829, -0.10474107414484024, 0.09766335785388947, 0.016502566635608673, -0.10581766068935394, -0.2191290557384491, -0.10095275193452835, 0.09381274878978729, 0.058123111724853516, 0.0941794216632843, -0.0025956255849450827, 0.2262309193611145, -0.0690128356218338, 0.08464787900447845, 0.27244797348976135, -0.3110087811946869, -0.06204682216048241, 0.003115382045507431, 0.062032707035541534, 0.04376010596752167, -0.08253242820501328, -0.037422873079776764, 0.061025336384773254, 0.06917066127061844, 0.12343906611204147, -0.02113879844546318, -0.0333096869289875, -0.01811666041612625, -0.1473993957042694, -0.04512237757444382, 0.16906510293483734, 0.013635123148560524, -0.07460976392030716, -0.07006964087486267, -0.0700935572385788, -0.11553466320037842, -0.03701610863208771, -0.06005319952964783, 0.024691492319107056, -0.07838480919599533, -0.09106213599443436, -0.02342264913022518, -0.06270531564950943, -0.03550416976213455, -0.0512026771903038, 0.17377623915672302, 0.055536169558763504, 0.049914032220840454, -0.07807371765375137, 0.07911646366119385, -0.046351395547389984, -0.1285974532365799, -0.017186904326081276, -0.02009599842131138, 0.009644398465752602, -0.05833557993173599, -0.03608468174934387, -0.05039031431078911, -0.012172877788543701, 0.19013407826423645, -0.11604301631450653, 0.03741159290075302, 0.020054511725902557, 0.014545260928571224, -0.005241751670837402, 0.1698230504989624, -0.049223627895116806, 0.02419133298099041, 0.017215529456734657, 0.07593706995248795, 0.026382841169834137, -0.022408707067370415, -0.11242200434207916, -0.007064052391797304, 0.11832130700349808, 0.026761610060930252, -0.040248673409223557, 0.041855812072753906, -0.02073734439909458, -0.046279191970825195, 0.04707340896129608, -0.11334803700447083, 0.03619607537984848, -0.024475112557411194, -0.08638644963502884, 0.024236353114247322, -0.0006719715893268585, -0.011585507541894913, -0.04784350469708443, 0.05897897109389305, -0.10750173777341843, 0.03204413875937462, -0.07894282042980194, -0.10653629899024963, 0.02004607394337654, -0.10894925892353058, -0.0018548623193055391, -0.09197533875703812, -0.14471082389354706, 0.006253518629819155, 0.030955811962485313, -0.05349881947040558, -0.030853567644953728, -0.029356731101870537, -0.07512738555669785, 0.0040463791228830814, -0.010720531456172466, 0.015385432168841362, -0.03507460653781891, 0.07795532792806625, 0.06438208371400833, 0.07988894730806351, -0.036637548357248306, 0.04111693054437637, -0.11104508489370346, 0.03658684343099594, -0.22151629626750946, 0.0532694011926651, -0.07029316574335098, 0.04425838217139244, -0.10827615112066269, -0.13768045604228973, 0.07341479510068893, -0.038132816553115845, 0.1106347143650055, 0.13878044486045837, -0.13995826244354248, -0.04432953521609306, 0.16929495334625244, -0.08402355760335922, -0.11322060972452164, 0.13442915678024292, -0.05477425828576088, 0.01787211373448372, 0.07008762657642365, 0.18688829243183136, 0.015114307403564453, -0.11608908325433731, -0.03552825748920441, -0.011888417415320873, 0.03925975039601326, -0.052419357001781464, 0.06031451001763344, 0.027141962200403214, 0.042391639202833176, 0.012069420889019966, -0.002925252076238394, -0.013360485434532166, -0.12454324215650558, -0.0755920335650444, -0.059068433940410614, -0.09101876616477966, 0.09598608314990997, 0.029466964304447174, 0.0669846385717392, -0.10042767226696014, -0.09707588702440262, 0.08203751593828201, 0.08322812616825104, -0.05293062701821327, 0.026616226881742477, -0.13746634125709534, 0.13366709649562836, -0.09972000867128372, -0.03778829053044319, -0.21533609926700592, -0.0452013835310936, 0.03835505247116089, 0.014165922068059444, 0.04464937373995781, 0.013993254862725735, 0.06081649288535118, 0.07735932618379593, -0.022604454308748245, -0.019362501800060272, -0.06598284840583801, -0.010702898725867271, -0.12883605062961578, -0.14371870458126068, -0.06375543773174286, -0.04599449783563614, 0.060112446546554565, -0.15814565122127533, 0.02520962618291378, 0.018584970384836197, 0.1276068389415741, 0.011428777128458023, -0.027947228401899338, -0.012363673187792301, 0.0646401196718216, -0.018295366317033768, -0.07542243599891663, 0.05274580419063568, 0.0023584654554724693, -0.08776634931564331, -0.08649705350399017, -0.1299140602350235, 0.12408684194087982, 0.10966097563505173, 0.005134599283337593, -0.07533600181341171, 0.002831381047144532, -0.07506682723760605, -0.014382652938365936, -0.05087573453783989, 0.03402669355273247, 0.2190069556236267, 0.014410837553441525, 0.14459405839443207, -0.08733177185058594, -0.05628376826643944, 0.03651189059019089, -0.030042355880141258, -0.010371820069849491, 0.08439674228429794, 0.030930375680327415, -0.12146168947219849, 0.10718461871147156, 0.11462022364139557, -0.06389530003070831, 0.08779355138540268, -0.05631178244948387, -0.07001283019781113, -0.04125839099287987, -0.02466856874525547, -0.0029342821799218655, 0.12691821157932281, -0.11756148934364319, -0.03004525415599346, 0.05033399164676666, 0.004875210113823414, 0.007627365179359913, -0.17963916063308716, -0.018573971465229988, 0.02412322908639908, -0.007105635944753885, -0.07540324330329895, 0.004033748526126146, 0.03410760685801506, 0.10098737478256226, 0.026414459571242332, -0.06166202202439308, 0.03965643048286438, 0.01726614311337471, -0.06112029403448105, 0.18413394689559937, -0.10484959185123444, -0.14828723669052124, -0.08433350175619125, -0.03477618843317032, -0.0659618079662323, -0.008506464771926403, 0.05553193762898445, -0.09697128087282181, -0.04002172127366066, -0.06898635625839233, -0.01814378798007965, -0.03358377516269684, 0.011067286133766174, 0.10296545177698135, -0.03468313068151474, 0.07709599286317825, -0.12533262372016907, -0.025831522420048714, -0.03703936189413071, -0.026492970064282417, 0.044200919568538666, 0.03433454409241676, 0.11545658111572266, 0.12074284255504608, -0.06155486777424812, 0.047797296196222305, -0.03470879793167114, 0.30651339888572693, -0.05929180607199669, -0.03354427218437195, 0.158450186252594, 0.0060304999351501465, 0.07536483556032181, 0.08727488666772842, 0.05836986377835274, -0.0994449183344841, -0.0007933140732347965, 0.0037772441282868385, -0.037853218615055084, -0.23493672907352448, -0.03564998507499695, -0.017255961894989014, -0.028254026547074318, 0.10449212789535522, 0.013824822381138802, 0.0394839346408844, 0.07591623812913895, 0.0072249784134328365, 0.08304719626903534, -0.06100599095225334, 0.11254733800888062, 0.11810389161109924, 0.03804830461740494, 0.12200139462947845, -0.034886434674263, -0.07265238463878632, 0.03962704539299011, -0.031471092253923416, 0.2563975155353546, -0.0029355252627283335, 0.09562867879867554, 0.04019680619239807, 0.1609010547399521, 0.007890925742685795, 0.08964591473340988, -0.01366937905550003, -0.019392352551221848, -0.007639882154762745, -0.04829324781894684, -0.06196831911802292, -0.004592290148139, -0.03607151284813881, 0.06786858290433884, -0.12941277027130127, 0.05080967769026756, 0.01941109076142311, 0.2504429519176483, 0.02993147075176239, -0.35981589555740356, -0.12131378054618835, -0.013684969395399094, -0.029202425852417946, -0.061327047646045685, 0.01961236633360386, 0.088455431163311, -0.10485497862100601, 0.03192731365561485, -0.07576137781143188, 0.10069511085748672, -0.028685791417956352, 0.03115217015147209, 0.014687330462038517, 0.1186612918972969, -0.00022383584291674197, 0.07261960208415985, -0.23363767564296722, 0.23800943791866302, 0.02963573858141899, 0.09928041696548462, -0.08848042786121368, 0.0006318351370282471, -0.023187914863228798, 0.10918482393026352, 0.10642307996749878, -0.005781724117696285, -0.04022325575351715, -0.17412349581718445, -0.08146384358406067, 0.017494481056928635, 0.09598741680383682, 0.018748227506875992, 0.11133407801389694, -0.013488655909895897, 0.018274696543812752, 0.02646196447312832, -0.037733256816864014, -0.1056593582034111, -0.09837887436151505, 0.03553704544901848, 0.030940989032387733, -0.05402953922748566, -0.06883232295513153, -0.12366290390491486, -0.03698309510946274, 0.15054598450660706, 0.04874749854207039, -0.06965745240449905, -0.12995630502700806, 0.06478682905435562, 0.1110588014125824, -0.07307134568691254, 0.012917744927108288, -0.005386942531913519, 0.11402306705713272, -0.012386246584355831, -0.08819925785064697, 0.07176806777715683, -0.06931772083044052, -0.15321139991283417, -0.019712964072823524, 0.11074908822774887, 0.06326816231012344, 0.06602394580841064, 0.011340194381773472, 0.022040799260139465, -0.01772451028227806, -0.08636745810508728, 0.01200300082564354, 0.03812067583203316, 0.10961323976516724, -0.0037691094912588596, -0.008088557049632072, 0.06079507991671562, -0.0672171413898468, 0.03737341985106468, 0.15773825347423553, 0.21972469985485077, -0.09252112358808517, 0.07291530817747116, 0.06872342526912689, -0.040819987654685974, -0.16155892610549927, 0.029970459640026093, 0.06614790111780167, 0.005190686322748661, 0.0009222908993251622, -0.1510581523180008, 0.09531722217798233, 0.0719270408153534, -0.009691058658063412, 0.03956225886940956, -0.3003876805305481, -0.10106629133224487, 0.10027233511209488, 0.11328063905239105, 0.14576928317546844, -0.12089318782091141, -0.02525998465716839, -0.01040996890515089, -0.14808805286884308, 0.10230496525764465, -0.08747659623622894, 0.09695066511631012, -0.0407709926366806, 0.16085994243621826, 0.024694496765732765, -0.0582515150308609, 0.12009768933057785, 0.027762258425354958, 0.0920654684305191, -0.06880522519350052, -0.03411891683936119, 0.15613211691379547, -0.0667184591293335, 0.08743080496788025, -0.0012995266588404775, 0.05626193806529045, -0.1256784349679947, -0.014454485848546028, -0.07633952796459198, 0.04748973622918129, -0.040521711111068726, -0.046718940138816833, -0.05238937959074974, 0.029814530164003372, 0.04296720400452614, -0.0263314638286829, 0.1443767547607422, 0.06497303396463394, 0.13215650618076324, 0.10146354883909225, 0.1296021193265915, -0.10676542669534683, -0.09109240025281906, 0.016986310482025146, -0.01820150949060917, 0.06846044957637787, -0.11617839336395264, 0.018155181780457497, 0.15312615036964417, 0.039134517312049866, 0.11135297268629074, 0.06829814612865448, -0.06438540667295456, 0.015836521983146667, 0.016460929065942764, -0.17109189927577972, -0.15569303929805756, -0.0024609605316072702, 0.040841687470674515, -0.13128487765789032, 0.09511961042881012, 0.09916234761476517, -0.08121468871831894, -0.04664384573698044, -0.017647631466388702, -0.015537886880338192, -0.012126259505748749, 0.16343770921230316, 0.06174853444099426, 0.07514795660972595, -0.11120883375406265, 0.0782356783747673, 0.06359988451004028, -0.08194202184677124, 0.05627282336354256, 0.05743302032351494, -0.0983603447675705, -0.025110265240073204, 0.05645470693707466, 0.1692047417163849, -0.0897742211818695, -0.026884572580456734, -0.12965117394924164, -0.0930682048201561, 0.07110972702503204, 0.17640617489814758, 0.07790535688400269, -0.03186081349849701, -0.04519864171743393, 0.018123148009181023, -0.16040417551994324, 0.08979874104261398, 0.03623822703957558, 0.06600173562765121, -0.1458890438079834, 0.13826458156108856, -0.01348023395985365, 0.09205173701047897, -0.029944607987999916, 0.005435056518763304, -0.1076556071639061, 0.010646708309650421, -0.1598055213689804, -0.01668854057788849, -0.060602616518735886, 0.001294052810408175, -0.025370348244905472, -0.05356859043240547, -0.04239482060074806, 0.04662984982132912, -0.08802337944507599, -0.010235413908958435, 0.04446113854646683, 0.003834289498627186, -0.13975214958190918, -0.008976796641945839, -0.0027601555921137333, -0.07138504087924957, 0.08376792073249817, 0.07584337890148163, -0.013954635709524155, 0.00865698792040348, -0.019008509814739227, -0.02887219563126564, 0.024132147431373596, 0.01300074253231287, 0.11923792958259583, -0.07713138312101364, -0.02087589167058468, -0.00256345490925014, 0.03576859086751938, 0.033608581870794296, 0.06675190478563309, -0.10265257209539413, -0.035551510751247406, -0.001128505333326757, -0.05532321333885193, -0.07866483181715012, 0.06195364147424698, 0.11472298204898834, 0.022822361439466476, 0.15290436148643494, -0.06175330653786659, 0.06683219224214554, -0.16610868275165558, -0.031011024489998817, 0.0013024181826040149, -0.07342507690191269, -0.04638872295618057, -0.03974561765789986, 0.07294844835996628, -0.069546177983284, 0.1149602010846138, -0.004614647943526506, 0.09820378571748734, 0.03352729603648186, -0.03708178550004959, -0.04204994812607765, -0.01278035156428814, 0.15957039594650269, 0.02860177308320999, -0.03492225706577301, 0.08943773061037064, 0.021394900977611542, 0.07308758050203323, 0.09829479455947876, 0.1944517344236374, 0.1139134019613266, 0.01851532980799675, 0.07470504194498062, 0.0435343012213707, -0.025490283966064453, -0.20076684653759003, -0.006181175354868174, 0.006392240524291992, 0.09106814861297607, -0.02206818200647831, 0.17320683598518372, 0.11280245333909988, -0.11516893655061722, 0.05252739042043686, -0.057665687054395676, -0.10511697828769684, -0.11207513511180878, -0.0900212973356247, -0.06883911043405533, -0.13371394574642181, 0.023637305945158005, -0.13672785460948944, 0.017487356439232826, 0.04450596868991852, 0.01315749529749155, -0.023366376757621765, 0.16111795604228973, -0.005352434236556292, 0.034207407385110855, 0.06221303716301918, 0.0019792867824435234, -0.02469668723642826, -0.026513831689953804, -0.029618175700306892, 0.025340432301163673, -0.009335009381175041, 0.07646328955888748, -0.018447615206241608, -0.009602234698832035, 0.02114568091928959, -0.021930528804659843, -0.09679368138313293, 0.012470961548388004, 0.034363895654678345, 0.0657031238079071, 0.03325597196817398, 0.015201108530163765, 0.0020816675387322903, -0.05482308194041252, 0.2031811624765396, -0.07466990500688553, -0.06195467710494995, -0.12251299619674683, 0.279879093170166, 0.035646721720695496, -0.03121574968099594, 0.05512085556983948, -0.08866076916456223, -0.017685027793049812, 0.20134049654006958, 0.14900332689285278, -0.032850995659828186, -0.014981723390519619, -0.02765616402029991, -0.015769800171256065, -0.060799721628427505, 0.082018181681633, 0.12076757848262787, 0.027126096189022064, -0.07983560860157013, -0.029441291466355324, -0.07358164340257645, -0.017690643668174744, -0.08899511396884918, 0.038366954773664474, 0.035992853343486786, 0.015548771247267723, -0.0586174838244915, 0.03955916687846184, -0.060560569167137146, -0.10813003033399582, 0.03239411488175392, -0.17456281185150146, -0.15235412120819092, 0.003399759065359831, 0.05891062319278717, -0.0006440808065235615, 0.08603335916996002, -0.015686845406889915, -0.0033804194536060095, 0.12489918619394302, -0.005691363476216793, -0.07472037523984909, -0.06126156449317932, 0.1348992884159088, -0.11912516504526138, 0.18333494663238525, -0.004152887035161257, 0.104251928627491, 0.12649808824062347, 0.028684085234999657, -0.13483522832393646, 0.026998115703463554, 0.052571073174476624, -0.03426850959658623, -0.015272337011992931, 0.1584072709083557, -0.007267709821462631, 0.054042667150497437, 0.027835864573717117, -0.1654953509569168, -0.04310823604464531, -0.02993229404091835, 0.014033780433237553, -0.06036237254738808, -0.0055208029225468636, -0.05640485882759094, 0.13550224900245667, 0.19209660589694977, -0.04742174223065376, -0.007382064592093229, -0.06235097348690033, 0.03008194826543331, 0.0720757246017456, 0.0001828142412705347, -0.050304755568504333, -0.23718851804733276, 0.010550931096076965, 0.05144874006509781, -0.00039068752084858716, -0.2348301261663437, -0.08169503509998322, 0.026268480345606804, -0.05515012517571449, -0.07741531729698181, 0.06445730477571487, 0.05951142683625221, 0.046166881918907166, -0.051601212471723557, -0.0339188277721405, -0.0839768573641777, 0.14498931169509888, -0.14945799112319946, -0.04033840447664261 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ai-club-inductions-21-nlp-roBERTa-base-squad-v2 This model was trained from scratch on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cpu - Datasets 1.14.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "ai-club-inductions-21-nlp-roBERTa-base-squad-v2", "results": []}]}
question-answering
AyushPJ/ai-club-inductions-21-nlp-roBERTa-base-squad-v2
[ "transformers", "pytorch", "roberta", "question-answering", "generated_from_trainer", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #roberta #question-answering #generated_from_trainer #endpoints_compatible #region-us
# ai-club-inductions-21-nlp-roBERTa-base-squad-v2 This model was trained from scratch on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cpu - Datasets 1.14.0 - Tokenizers 0.10.3
[ "# ai-club-inductions-21-nlp-roBERTa-base-squad-v2\n\nThis model was trained from scratch on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 5", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #roberta #question-answering #generated_from_trainer #endpoints_compatible #region-us \n", "# ai-club-inductions-21-nlp-roBERTa-base-squad-v2\n\nThis model was trained from scratch on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 5", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ 37, 39, 6, 12, 8, 3, 90, 33 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #question-answering #generated_from_trainer #endpoints_compatible #region-us \n# ai-club-inductions-21-nlp-roBERTa-base-squad-v2\n\nThis model was trained from scratch on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 5### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ -0.12262574583292007, 0.11893592774868011, -0.0008845579577609897, 0.08377688378095627, 0.1274728924036026, 0.019584007561206818, 0.11258997768163681, 0.128732368350029, -0.09086467325687408, 0.07562503963708878, 0.10681810230016708, 0.07517492771148682, 0.04055452719330788, 0.15725165605545044, -0.0317697636783123, -0.19977235794067383, 0.05390666797757149, 0.05113222822546959, -0.04388023540377617, 0.1156042143702507, 0.09390133619308472, -0.13146361708641052, 0.08760994672775269, 0.021455220878124237, -0.21223849058151245, -0.005762892309576273, -0.026467010378837585, -0.09250123798847198, 0.08898759633302689, -0.01628006249666214, 0.10909189283847809, 0.00806707888841629, 0.0911700502038002, -0.12174678593873978, 0.00591759430244565, 0.04531410336494446, 0.02427052892744541, 0.11284837871789932, 0.01661762408912182, -0.009021669626235962, 0.05768671631813049, -0.1128302812576294, 0.09996189177036285, 0.027088157832622528, -0.1017245352268219, -0.20130810141563416, -0.09600841999053955, 0.09901370853185654, 0.06821171194314957, 0.08278930932283401, 0.0012600277550518513, 0.21187953650951385, -0.09545236080884933, 0.05894877761602402, 0.22359992563724518, -0.30827832221984863, -0.056728579103946686, 0.051134075969457626, 0.05103736370801926, 0.042238567024469376, -0.08947673439979553, -0.02825215272605419, 0.055460941046476364, 0.0497339591383934, 0.09810903668403625, -0.027064895257353783, -0.09142089635133743, -0.00913575105369091, -0.1354617476463318, -0.05373349040746689, 0.2033509761095047, 0.019581658765673637, -0.06612757593393326, -0.07364220172166824, -0.0727459043264389, -0.12034603953361511, -0.02737104520201683, -0.039735086262226105, 0.020145276561379433, -0.06806031614542007, -0.1026747077703476, -0.021629123017191887, -0.04797765985131264, -0.07075148820877075, -0.0331178717315197, 0.13227520883083344, 0.058894988149404526, 0.05576976388692856, -0.0424058772623539, 0.11009316891431808, -0.041154250502586365, -0.10429748147726059, -0.03058564104139805, -0.014993474818766117, -0.04019446298480034, -0.05012348294258118, -0.033741749823093414, -0.017345042899250984, -0.005051013082265854, 0.19756847620010376, -0.0540601946413517, 0.03301957994699478, 0.029532240703701973, 0.006174384616315365, 0.0008840978844091296, 0.15547236800193787, -0.07221865653991699, 0.00523768924176693, 0.025126313790678978, 0.06363271921873093, 0.011712136678397655, -0.018072685226798058, -0.09357345849275589, -0.03220997005701065, 0.10949359834194183, 0.05066889151930809, -0.0423365943133831, 0.04178719222545624, -0.017601236701011658, -0.05612148717045784, 0.05241822451353073, -0.11231798678636551, 0.0374419130384922, -0.013298812322318554, -0.08856330811977386, 0.02639319933950901, 0.014318102039396763, -0.020476622506976128, -0.05565161630511284, 0.025965603068470955, -0.1130337342619896, 0.008680827915668488, -0.08337471634149551, -0.09152039140462875, 0.028322525322437286, -0.11442180722951889, -0.013553142547607422, -0.08570777624845505, -0.16229642927646637, 0.013447089120745659, 0.020973922684788704, -0.06532715260982513, -0.039741531014442444, -0.022691864520311356, -0.06747308373451233, 0.014461721293628216, -0.008861051872372627, 0.04477095976471901, -0.03258470073342323, 0.06924452632665634, 0.05693931505084038, 0.06972106546163559, -0.011163179762661457, 0.04614165052771568, -0.1098981499671936, 0.0390273854136467, -0.21414706110954285, 0.0701150968670845, -0.07219871133565903, 0.02823130041360855, -0.12093684077262878, -0.10848232358694077, 0.0718110054731369, -0.03359610587358475, 0.08626718819141388, 0.14213736355304718, -0.10762641578912735, -0.03191487863659859, 0.16461291909217834, -0.08653996139764786, -0.12655572593212128, 0.11483179032802582, -0.04609208181500435, 0.012463831342756748, 0.07046503573656082, 0.15667983889579773, 0.050318602472543716, -0.10941438376903534, -0.05122387781739235, 0.006157896015793085, 0.0492120236158371, -0.07768244296312332, 0.08072322607040405, 0.01656489633023739, 0.008350979536771774, 0.021255994215607643, -0.04569403827190399, -0.01166551373898983, -0.10522913932800293, -0.08263170719146729, -0.06975600868463516, -0.09275020658969879, 0.06977005302906036, 0.016200916841626167, 0.055280379951000214, -0.07612836360931396, -0.09218895435333252, 0.06524007767438889, 0.13061530888080597, -0.06664394587278366, 0.017460716888308525, -0.12610678374767303, 0.11948039382696152, -0.1286265105009079, -0.014873974025249481, -0.20943474769592285, -0.06858653575181961, 0.04330439120531082, -0.01867307722568512, 0.03887603059411049, 0.01746063306927681, 0.050695598125457764, 0.07069049030542374, -0.02334672398865223, -0.020982077345252037, -0.07605449110269547, -0.019304171204566956, -0.1081194058060646, -0.15273411571979523, -0.0763821005821228, -0.036554452031850815, 0.11329121142625809, -0.16912707686424255, 0.018942764028906822, 0.025658685714006424, 0.1343664973974228, 0.019027963280677795, -0.028473468497395515, 0.0031161976512521505, 0.05481502786278725, -0.01852731592953205, -0.08114328235387802, 0.03749607503414154, -0.006362877320498228, -0.08803832530975342, -0.06916864961385727, -0.11251527070999146, 0.0889301672577858, 0.11060448735952377, 0.03128615766763687, -0.07533460110425949, 0.019526734948158264, -0.06317045539617538, -0.022418687120079994, -0.0630354955792427, 0.01347413007169962, 0.17441634833812714, 0.002399433869868517, 0.14435091614723206, -0.08409294486045837, -0.0454624742269516, 0.03441045060753822, -0.018928276374936104, -0.005166064016520977, 0.07982183247804642, 0.03924033045768738, -0.09362158179283142, 0.11655089259147644, 0.10121091455221176, -0.055604487657547, 0.08876996487379074, -0.05321570858359337, -0.08967713266611099, -0.04702497646212578, -0.008800959214568138, 0.0020553688518702984, 0.14380180835723877, -0.11757906526327133, -0.02815055102109909, 0.05497225001454353, 0.014776922762393951, 0.013104060664772987, -0.16909900307655334, -0.008579575456678867, 0.013610648922622204, -0.02700425125658512, -0.044029977172613144, 0.004254661500453949, 0.03388434275984764, 0.10070677101612091, 0.026210704818367958, -0.04159540310502052, 0.04419746994972229, 0.01284739375114441, -0.06776227802038193, 0.1941862255334854, -0.09762407094240189, -0.1459699273109436, -0.09708636999130249, -0.0036476287059485912, -0.07076646387577057, -0.0147636029869318, 0.02129172906279564, -0.08731379359960556, -0.0473051592707634, -0.07750776410102844, -0.010004667565226555, -0.04031449928879738, 0.011969951912760735, 0.06412404030561447, -0.005900803953409195, 0.08025187999010086, -0.12375300377607346, -0.0040329694747924805, -0.03269946202635765, -0.08318214118480682, 0.02727184072136879, 0.043180327862501144, 0.1254507303237915, 0.13555318117141724, -0.04832172766327858, 0.044958025217056274, -0.023545945063233376, 0.2615821361541748, -0.06524798274040222, -0.03976409137248993, 0.12420489639043808, 0.007529493886977434, 0.04422175511717796, 0.09722823649644852, 0.046216391026973724, -0.08703628182411194, -0.004380333237349987, 0.0005495535442605615, -0.03664444386959076, -0.24799296259880066, -0.05035536736249924, -0.03122437745332718, -0.010014468804001808, 0.10923764109611511, 0.02702275477349758, 0.04914402589201927, 0.09090356528759003, -0.0034610056318342686, 0.10408590734004974, -0.06167612969875336, 0.10051377862691879, 0.1015445739030838, 0.04890785366296768, 0.1013910323381424, -0.054760295897722244, -0.06539123505353928, 0.042839571833610535, -0.00026576066738925874, 0.25308161973953247, -0.026068536564707756, 0.12658585608005524, 0.02655152976512909, 0.18860463798046112, 0.016191890463232994, 0.08937999606132507, -0.008035712875425816, -0.009828823618590832, -0.003839571727439761, -0.04965652525424957, -0.05357293039560318, 0.0009443223825655878, -0.011524374596774578, 0.07171110063791275, -0.12252975255250931, 0.015985343605279922, 0.025932295247912407, 0.2635839879512787, 0.05597972497344017, -0.3556210398674011, -0.1054057702422142, -0.003034197958186269, -0.010634119622409344, -0.06365495175123215, 0.009652979671955109, 0.09307222813367844, -0.12318042665719986, 0.024253344163298607, -0.06971726566553116, 0.0994991660118103, -0.05591466277837753, -0.0019094389863312244, 0.01742391847074032, 0.13260193169116974, 0.0015093362890183926, 0.08561254292726517, -0.19791291654109955, 0.21198196709156036, 0.03837864100933075, 0.09792570769786835, -0.09084325283765793, 0.0038765459321439266, -0.020047524943947792, 0.09013905376195908, 0.09836295992136002, -0.003956153057515621, -0.05439581722021103, -0.19952556490898132, -0.10985170304775238, 0.03176906332373619, 0.07632264494895935, -0.011324218474328518, 0.10855238139629364, -0.02156231179833412, 0.008777251467108727, 0.019732128828763962, -0.043370023369789124, -0.11229533702135086, -0.09827949851751328, 0.03059981018304825, 0.041717469692230225, -0.02439802512526512, -0.06126172095537186, -0.10250868648290634, 0.0050604441203176975, 0.14699570834636688, 0.014723539352416992, -0.046755608171224594, -0.13299806416034698, 0.06001821905374527, 0.137606680393219, -0.07428254187107086, 0.022363809868693352, -0.0070418864488601685, 0.11821961402893066, 0.018141407519578934, -0.09553676843643188, 0.07430166751146317, -0.07093922048807144, -0.15738026797771454, -0.0199043657630682, 0.10287903249263763, 0.05493176355957985, 0.06265582144260406, 0.012322389520704746, 0.023779025301337242, -0.009254546836018562, -0.0879349410533905, 0.009259003214538097, 0.060432709753513336, 0.08011505007743835, 0.06302469968795776, -0.03127652406692505, 0.03216879442334175, -0.033440250903367996, 0.026103567332029343, 0.15588392317295074, 0.20221585035324097, -0.09425006806850433, 0.07007135450839996, 0.09602446109056473, -0.0444520004093647, -0.1850166618824005, 0.040330104529857635, 0.04303120821714401, 0.021708594635128975, -0.002159625058993697, -0.1618398129940033, 0.1135796457529068, 0.0836556926369667, -0.015394317917525768, 0.04586603119969368, -0.31747397780418396, -0.09658646583557129, 0.11058390885591507, 0.09374645352363586, 0.11722715198993683, -0.11260726302862167, -0.0326116681098938, -0.016391992568969727, -0.1356222927570343, 0.116018146276474, -0.1302003562450409, 0.1159779503941536, -0.049752023071050644, 0.1346564143896103, 0.017079319804906845, -0.05526166036725044, 0.13217584788799286, 0.050034333020448685, 0.08167537301778793, -0.05398871377110481, -0.010310451500117779, 0.1766485720872879, -0.06742832064628601, 0.1034177616238594, -0.0016325770411640406, 0.05975606292486191, -0.1458095908164978, -0.008419732563197613, -0.06819064915180206, 0.04944681376218796, -0.03472774475812912, -0.04279357194900513, -0.06572891026735306, 0.03609044477343559, 0.034146782010793686, -0.03303322196006775, 0.10815390199422836, 0.07663799822330475, 0.11710549890995026, 0.09710705280303955, 0.10367603600025177, -0.049755945801734924, -0.11788161098957062, 0.02138298936188221, -0.021816592663526535, 0.08406403660774231, -0.12234465032815933, 0.02281162329018116, 0.13183705508708954, 0.038803063333034515, 0.12056215852499008, 0.058828044682741165, -0.07574547827243805, 0.018557215109467506, 0.014030414633452892, -0.15361572802066803, -0.11400875449180603, 0.005579126998782158, -0.014508143067359924, -0.14417365193367004, 0.0817563384771347, 0.10407928377389908, -0.09279145300388336, -0.03945521265268326, -0.013185715302824974, -0.01487644575536251, -0.027906225994229317, 0.1817956119775772, 0.07494298368692398, 0.07244228571653366, -0.12114287167787552, 0.10798196494579315, 0.07109905779361725, -0.08943405002355576, 0.05114598199725151, 0.045857932418584824, -0.10645157843828201, -0.035066861659288406, 0.04967112839221954, 0.16135792434215546, -0.08597393333911896, -0.034505125135183334, -0.11211436241865158, -0.09840398281812668, 0.06958109140396118, 0.12462147325277328, 0.07552856206893921, -0.0046999393962323666, -0.039428792893886566, 0.011355324648320675, -0.16483744978904724, 0.0865325778722763, 0.04111244156956673, 0.06161879748106003, -0.14606039226055145, 0.10877873003482819, 0.002837521955370903, 0.08340373635292053, -0.030718909576535225, -0.00008054752834141254, -0.10718173533678055, 0.004925955086946487, -0.1521265059709549, -0.01487649418413639, -0.060085996985435486, -0.00028478680178523064, -0.02181529812514782, -0.04418458417057991, -0.04882010817527771, 0.03982303664088249, -0.08726753294467926, -0.025460422039031982, 0.027284665033221245, 0.019831281155347824, -0.14276520907878876, -0.008160963654518127, 0.006505604367703199, -0.08398590236902237, 0.08562366664409637, 0.07948201894760132, 0.003377008717507124, 0.0034114618320018053, -0.03418254852294922, -0.03940876945853233, 0.016892695799469948, 0.014853158965706825, 0.11619367450475693, -0.0804576501250267, -0.002757707145065069, -0.004039627965539694, 0.04243728518486023, 0.031809594482183456, 0.07154525816440582, -0.10932447016239166, -0.0485912561416626, -0.044739335775375366, -0.04976833239197731, -0.07424720376729965, 0.06927679479122162, 0.13253259658813477, 0.04203276336193085, 0.14487411081790924, -0.060779597610235214, 0.05997684225440025, -0.17925414443016052, -0.03811082988977432, -0.011495104990899563, -0.05431937426328659, -0.08446669578552246, -0.023869041353464127, 0.06989701837301254, -0.0672801062464714, 0.13744628429412842, -0.021370762959122658, 0.08856581151485443, 0.02510250359773636, -0.01458432711660862, -0.02145782671868801, 0.006308740004897118, 0.17159652709960938, 0.05392121151089668, -0.022229928523302078, 0.10372130572795868, 0.024031512439250946, 0.07349127531051636, 0.08105912059545517, 0.15129497647285461, 0.09341789782047272, 0.04001405090093613, 0.09051595628261566, 0.05914675444364548, -0.05383634567260742, -0.14780643582344055, 0.03111354634165764, -0.015403621830046177, 0.09629686176776886, -0.01588612049818039, 0.16208045184612274, 0.11235933750867844, -0.1306961327791214, 0.04330452159047127, -0.04851838946342468, -0.11338111758232117, -0.11032676696777344, -0.0800471156835556, -0.07880426198244095, -0.1457136869430542, 0.014382630586624146, -0.13347145915031433, -0.006637387443333864, 0.08272993564605713, 0.010254105553030968, -0.007146542426198721, 0.17035028338432312, 0.012185477651655674, 0.02647067792713642, 0.055158473551273346, 0.01475248858332634, -0.015344151295721531, -0.02244408428668976, -0.030395520851016045, 0.038514185696840286, -0.00661626597866416, 0.0869738906621933, -0.018479837104678154, -0.0002948923211079091, 0.054822225123643875, -0.02084648422896862, -0.09691010415554047, 0.011267263442277908, 0.029966501519083977, 0.05282013490796089, 0.020088257268071175, 0.030919983983039856, -0.009840292856097221, -0.05425535514950752, 0.20589563250541687, -0.0706513524055481, -0.03563244640827179, -0.13834017515182495, 0.22444231808185577, 0.028403889387845993, -0.036839473992586136, 0.06973547488451004, -0.09446658939123154, -0.018306642770767212, 0.1878339946269989, 0.18566057085990906, -0.0404798649251461, -0.02176492288708687, -0.027068940922617912, -0.027049517259001732, -0.05232363939285278, 0.09710178524255753, 0.09603612869977951, 0.03250985965132713, -0.07630720734596252, -0.011919227428734303, -0.051999785006046295, -0.0383947528898716, -0.07621442526578903, 0.03569665178656578, 0.020808646455407143, 0.006600330118089914, -0.05071389675140381, 0.04336022585630417, -0.04940326511859894, -0.10220456123352051, 0.06628682464361191, -0.16052819788455963, -0.16247297823429108, -0.0161859430372715, 0.03137008845806122, 0.013280141167342663, 0.07194066047668457, -0.030117560178041458, 0.0070061408914625645, 0.13158941268920898, -0.007432152982801199, -0.09483321011066437, -0.11471546441316605, 0.13296620547771454, -0.06057363748550415, 0.1997159719467163, -0.007337300572544336, 0.09267912805080414, 0.11910019814968109, 0.015949292108416557, -0.15589715540409088, 0.0012550167739391327, 0.0592329241335392, -0.03664929419755936, -0.005380805116146803, 0.15622250735759735, -0.021826839074492455, 0.07908553630113602, 0.03197307512164116, -0.14588014781475067, -0.04668647423386574, -0.04956858232617378, 0.03584311529994011, -0.07892415672540665, -0.02423865720629692, -0.054360318928956985, 0.14577950537204742, 0.1957826018333435, -0.0605560727417469, -0.006521948613226414, -0.06431557238101959, 0.03188098222017288, 0.060923982411623, 0.0008272401173599064, -0.02959071472287178, -0.22048261761665344, 0.014068345539271832, 0.04639424756169319, 0.012081144377589226, -0.23279806971549988, -0.09064681828022003, 0.025830570608377457, -0.06536950170993805, -0.08083346486091614, 0.08717001974582672, 0.04679599031805992, 0.05165218934416771, -0.046079110354185104, -0.041109442710876465, -0.07221127301454544, 0.13557502627372742, -0.15719230473041534, -0.05629628151655197 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # ai-club-inductions-21-nlp-roBERTa This model was trained from scratch on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cpu - Datasets 1.14.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "ai-club-inductions-21-nlp-roBERTa", "results": []}]}
question-answering
AyushPJ/ai-club-inductions-21-nlp-roBERTa
[ "transformers", "pytorch", "roberta", "question-answering", "generated_from_trainer", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #roberta #question-answering #generated_from_trainer #endpoints_compatible #region-us
# ai-club-inductions-21-nlp-roBERTa This model was trained from scratch on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cpu - Datasets 1.14.0 - Tokenizers 0.10.3
[ "# ai-club-inductions-21-nlp-roBERTa\n\nThis model was trained from scratch on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #roberta #question-answering #generated_from_trainer #endpoints_compatible #region-us \n", "# ai-club-inductions-21-nlp-roBERTa\n\nThis model was trained from scratch on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ 37, 31, 6, 12, 8, 3, 90, 33 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #question-answering #generated_from_trainer #endpoints_compatible #region-us \n# ai-club-inductions-21-nlp-roBERTa\n\nThis model was trained from scratch on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cpu\n- Datasets 1.14.0\n- Tokenizers 0.10.3" ]
[ -0.08873303979635239, 0.07693744450807571, -0.0022627969738096, 0.08932628482580185, 0.13505297899246216, -0.01782067120075226, 0.12793327867984772, 0.12228015810251236, -0.10462063550949097, 0.06403876841068268, 0.12388047575950623, 0.12414856255054474, 0.028676973655819893, 0.1534348577260971, -0.03745657578110695, -0.2365981787443161, 0.04455597326159477, 0.05909438803792, -0.010947571136057377, 0.12273827195167542, 0.09535272419452667, -0.14561666548252106, 0.08809018135070801, 0.04433140531182289, -0.19940969347953796, 0.010098160244524479, -0.01652379520237446, -0.09662267565727234, 0.09829241037368774, -0.01886650361120701, 0.10303359478712082, 0.011343097314238548, 0.08556563407182693, -0.1307489275932312, 0.009166009724140167, 0.04083986207842827, 0.028235159814357758, 0.10254126042127609, 0.023249683901667595, -0.022302325814962387, 0.08551367372274399, -0.10139257460832596, 0.10069403052330017, 0.025508783757686615, -0.11077409237623215, -0.23101678490638733, -0.09773223847150803, 0.083358533680439, 0.05788347125053406, 0.09344105422496796, -0.0016613524639979005, 0.23783980309963226, -0.07352778315544128, 0.08503830432891846, 0.275518000125885, -0.312637597322464, -0.06363631784915924, 0.013400843366980553, 0.06317884474992752, 0.056490104645490646, -0.08747982233762741, -0.026885658502578735, 0.05711446702480316, 0.06430088728666306, 0.12251292169094086, -0.02734137512743473, -0.0384991429746151, -0.013154475018382072, -0.141809344291687, -0.0523223839700222, 0.16451993584632874, 0.022856280207633972, -0.07208888977766037, -0.07071026414632797, -0.06870981305837631, -0.11630494892597198, -0.039859335869550705, -0.05693601816892624, 0.027340857312083244, -0.07297279685735703, -0.10833161324262619, -0.02271355502307415, -0.05397813767194748, -0.043730538338422775, -0.04033086076378822, 0.18755462765693665, 0.05598139390349388, 0.05090858414769173, -0.08242462575435638, 0.07972578704357147, -0.04062865301966667, -0.11464866995811462, -0.029971366748213768, -0.0167441014200449, 0.0017092369962483644, -0.051629550755023956, -0.03278178721666336, -0.05211428552865982, -0.0039022816345095634, 0.20303913950920105, -0.0927533358335495, 0.04025043547153473, 0.023368267342448235, 0.012469432316720486, -0.009488810785114765, 0.1756027489900589, -0.05410169064998627, 0.010419719852507114, 0.026510557159781456, 0.05923018977046013, 0.021355580538511276, -0.0189377274364233, -0.11051267385482788, -0.012458396144211292, 0.12239278852939606, 0.03691413626074791, -0.0408322848379612, 0.053442783653736115, -0.012487473897635937, -0.04776092618703842, 0.03441143035888672, -0.11157669872045517, 0.02848445624113083, -0.025874976068735123, -0.08658917993307114, 0.011876669712364674, 0.0033260383643209934, -0.013408605009317398, -0.041080232709646225, 0.04474121332168579, -0.1036539077758789, 0.021572621539235115, -0.07878435403108597, -0.0997905284166336, 0.016168516129255295, -0.10838501155376434, 0.0121614383533597, -0.10048643499612808, -0.1420823633670807, 0.0057092043571174145, 0.02791845239698887, -0.05504689738154411, -0.03296663612127304, -0.022669488564133644, -0.07254049926996231, -0.0021238394547253847, -0.010599777102470398, 0.034620240330696106, -0.035604190081357956, 0.0751800462603569, 0.06662138551473618, 0.07878148555755615, -0.017678314819931984, 0.04224928468465805, -0.11816894263029099, 0.031364817172288895, -0.2227105349302292, 0.054305654019117355, -0.06793417036533356, 0.054341867566108704, -0.1185019388794899, -0.1284179538488388, 0.060301005840301514, -0.031966179609298706, 0.0981869176030159, 0.1231931522488594, -0.09835272282361984, -0.05031586438417435, 0.18422052264213562, -0.08109727501869202, -0.11373982578516006, 0.12303958833217621, -0.05812530964612961, 0.025476375594735146, 0.0726432055234909, 0.18350112438201904, 0.020542412996292114, -0.10324237495660782, -0.02388504520058632, -0.01571659743785858, 0.017391588538885117, -0.056665509939193726, 0.06699637323617935, 0.02114059217274189, 0.03431366756558418, 0.015868769958615303, 0.0011725419899448752, -0.00752106960862875, -0.11902547627687454, -0.07366224378347397, -0.06169792264699936, -0.0896771177649498, 0.07962014526128769, 0.02635030634701252, 0.06937851011753082, -0.10229797661304474, -0.10185672342777252, 0.06366115063428879, 0.09022172540426254, -0.054179947823286057, 0.025621768087148666, -0.14194746315479279, 0.1258220225572586, -0.10798043012619019, -0.02975870668888092, -0.21557481586933136, -0.03536788001656532, 0.03444249555468559, -0.006301159504801035, 0.05235572159290314, 0.02045874111354351, 0.06633152067661285, 0.06354576349258423, -0.020250579342246056, -0.012838616967201233, -0.06194763630628586, -0.01265826914459467, -0.12075366824865341, -0.14489401876926422, -0.06497976183891296, -0.047305140644311905, 0.05592856928706169, -0.1545598804950714, 0.023626040667295456, 0.02924388460814953, 0.12738756835460663, 0.0148844625800848, -0.02552981860935688, -0.006382947787642479, 0.05812963843345642, -0.016079923138022423, -0.07910841703414917, 0.05186663195490837, -0.002303607063367963, -0.08675269782543182, -0.0732651874423027, -0.13317671418190002, 0.11193957924842834, 0.11259662359952927, 0.005144425202161074, -0.07465527951717377, 0.025222208350896835, -0.06601841747760773, -0.013305515050888062, -0.041562024503946304, 0.036309920251369476, 0.23331476747989655, 0.009527050890028477, 0.15107950568199158, -0.08495204895734787, -0.058405227959156036, 0.041389115154743195, -0.03380078449845314, 0.00017379903874825686, 0.08592167496681213, 0.022963184863328934, -0.13847850263118744, 0.10778316855430603, 0.10383006930351257, -0.05703941360116005, 0.09277822077274323, -0.05126950144767761, -0.06821324676275253, -0.037874896079301834, -0.02314041182398796, 0.000023336635422310792, 0.11390150338411331, -0.12469145655632019, -0.03194078430533409, 0.050672855228185654, 0.008496303111314774, 0.010846571996808052, -0.1702071875333786, -0.012572061270475388, 0.02875419706106186, -0.007324634585529566, -0.05605871230363846, 0.00909830629825592, 0.038073718547821045, 0.09994067251682281, 0.0381515771150589, -0.05651815980672836, 0.04194950312376022, 0.016248969361186028, -0.05904215946793556, 0.18632923066616058, -0.08773697167634964, -0.1422363668680191, -0.09053140878677368, -0.04515576735138893, -0.06773649901151657, -0.01016815472394228, 0.056803978979587555, -0.10068058967590332, -0.04056917876005173, -0.060635823756456375, -0.002507399767637253, -0.04564408212900162, 0.00582707766443491, 0.09518904983997345, -0.022411741316318512, 0.07868461310863495, -0.1281232088804245, -0.02285289578139782, -0.03893550485372543, -0.05070394650101662, 0.04001907631754875, 0.0321228913962841, 0.1231350526213646, 0.1201014444231987, -0.0676363855600357, 0.046198684722185135, -0.033230010420084, 0.2913546562194824, -0.07020074129104614, -0.04137660935521126, 0.15595555305480957, 0.013391966931521893, 0.0654124990105629, 0.08217039704322815, 0.05029641464352608, -0.09898688644170761, -0.0010115673067048192, 0.007901678793132305, -0.04700934514403343, -0.23802924156188965, -0.030506081879138947, -0.019965846091508865, -0.024918226525187492, 0.10191623121500015, 0.021543623879551888, 0.05687154829502106, 0.07862313836812973, 0.007246335037052631, 0.07781675457954407, -0.05103886127471924, 0.11960162967443466, 0.11398119479417801, 0.04319978505373001, 0.12089476734399796, -0.04416940361261368, -0.0784117579460144, 0.0418982058763504, -0.028535371646285057, 0.25393354892730713, 0.007383212447166443, 0.10322695225477219, 0.038488633930683136, 0.15661701560020447, 0.01521709281951189, 0.08043089509010315, -0.007767463568598032, -0.019908614456653595, -0.014631817117333412, -0.03959052264690399, -0.06962499767541885, -0.002359469886869192, -0.03544485941529274, 0.0605008527636528, -0.12142649292945862, 0.03598955273628235, 0.022988496348261833, 0.2615821361541748, 0.01815846748650074, -0.355892151594162, -0.1170954555273056, -0.012417849153280258, -0.03936780244112015, -0.057881101965904236, 0.025279806926846504, 0.07066898792982101, -0.1228678897023201, 0.0309764351695776, -0.07913557440042496, 0.10407376289367676, -0.03565860912203789, 0.025987159460783005, 0.00027695042081177235, 0.12857477366924286, -0.0012111702235415578, 0.07853977382183075, -0.21195247769355774, 0.23880356550216675, 0.026039693504571915, 0.09832867234945297, -0.08751098811626434, 0.0009387458558194339, -0.01719518192112446, 0.10112911462783813, 0.1062261164188385, -0.009909339249134064, -0.06800376623868942, -0.1769590824842453, -0.07079750299453735, 0.020981555804610252, 0.09742818027734756, 0.012513420544564724, 0.1158306673169136, -0.025906281545758247, 0.020080363377928734, 0.023673897609114647, -0.04917100816965103, -0.08836168795824051, -0.09214378148317337, 0.023333940654993057, 0.028875473886728287, -0.03958302363753319, -0.06793462485074997, -0.10732906311750412, -0.02486095391213894, 0.15483355522155762, 0.059657298028469086, -0.05578112229704857, -0.13144317269325256, 0.06168896332383156, 0.10488332062959671, -0.07952707260847092, 0.01593899354338646, -0.013102544471621513, 0.10962436348199844, -0.0011069545289501548, -0.0914137065410614, 0.07957053184509277, -0.06272468715906143, -0.14731371402740479, -0.010153586976230145, 0.09999805688858032, 0.05173295736312866, 0.059960756450891495, 0.012331120669841766, 0.025511732324957848, -0.018580438569188118, -0.0852610319852829, 0.01789304055273533, 0.025818217545747757, 0.08770749717950821, 0.006950758397579193, -0.000816031766589731, 0.058713894337415695, -0.07003209739923477, 0.0337340272963047, 0.1652328073978424, 0.22058996558189392, -0.09531181305646896, 0.07047116011381149, 0.07147694379091263, -0.03761061653494835, -0.16284841299057007, 0.022657984867691994, 0.048630110919475555, 0.0007480699568986893, 0.010122011415660381, -0.1478314995765686, 0.09260545670986176, 0.07346336543560028, -0.012066459283232689, 0.026110416278243065, -0.2933993637561798, -0.10050825774669647, 0.10484179109334946, 0.11752902716398239, 0.15212413668632507, -0.12255266308784485, -0.024616271257400513, -0.00928867980837822, -0.14856010675430298, 0.08120562136173248, -0.09920165687799454, 0.11181686073541641, -0.043168097734451294, 0.1562185138463974, 0.019786406308412552, -0.053763825446367264, 0.1129266545176506, 0.017049672082066536, 0.09558180719614029, -0.07169075310230255, -0.03297904133796692, 0.163812056183815, -0.06079188361763954, 0.07817293703556061, 0.00008667688962304965, 0.054203104227781296, -0.1289089024066925, -0.018343834206461906, -0.0799727514386177, 0.051710426807403564, -0.026054278016090393, -0.04630327224731445, -0.06296099722385406, 0.02788216806948185, 0.04395293816924095, -0.02574857510626316, 0.16096393764019012, 0.05311153456568718, 0.1303977072238922, 0.11177050322294235, 0.12550878524780273, -0.08827376365661621, -0.09453729540109634, 0.012152208015322685, -0.017218293622136116, 0.07009778171777725, -0.12114004790782928, 0.02085353620350361, 0.13483017683029175, 0.038098521530628204, 0.1037975326180458, 0.06425471603870392, -0.06812212616205215, 0.02802438475191593, 0.018280060961842537, -0.15333643555641174, -0.1582702100276947, 0.0021252876613289118, 0.012485165148973465, -0.1212138757109642, 0.09485023468732834, 0.1047343984246254, -0.08006399124860764, -0.047620877623558044, -0.02101912349462509, -0.020213080570101738, -0.02399951033294201, 0.15152078866958618, 0.06472773849964142, 0.07238271832466125, -0.11102620512247086, 0.08522789925336838, 0.06315885484218597, -0.0686543881893158, 0.05962293595075607, 0.057248055934906006, -0.09425656497478485, -0.026223696768283844, 0.060275766998529434, 0.1712641417980194, -0.09695089608430862, -0.02571353316307068, -0.13190671801567078, -0.09249506890773773, 0.06825094670057297, 0.17446017265319824, 0.07693314552307129, -0.03462929278612137, -0.04124264046549797, 0.022809715941548347, -0.164184108376503, 0.0802905261516571, 0.027974583208560944, 0.05799973011016846, -0.1479128897190094, 0.12171445786952972, -0.004167333710938692, 0.0878237783908844, -0.03314761444926262, -0.004216852132230997, -0.10773693025112152, 0.014814075082540512, -0.15909114480018616, -0.023754719644784927, -0.057670410722494125, 0.0025277382228523493, -0.02526620216667652, -0.05807758867740631, -0.052917420864105225, 0.04657624289393425, -0.09441603720188141, -0.011367462575435638, 0.04177827015519142, 0.012941018678247929, -0.1320052146911621, -0.007949398830533028, 0.00472598522901535, -0.06958342343568802, 0.07738117128610611, 0.07981480658054352, -0.011951851658523083, 0.013641182333230972, -0.02920142374932766, -0.025668269023299217, 0.02397499978542328, 0.0019479014445096254, 0.11479970812797546, -0.0812971293926239, -0.015459423884749413, -0.006139368750154972, 0.04306252300739288, 0.03190120309591293, 0.0582718588411808, -0.11202079802751541, -0.03291444107890129, -0.009318213909864426, -0.06061369180679321, -0.07759328931570053, 0.06191180273890495, 0.10912569612264633, 0.02047152817249298, 0.15311242640018463, -0.06257454305887222, 0.06802108138799667, -0.17123523354530334, -0.028716910630464554, -0.001214102259837091, -0.07916367799043655, -0.057693541049957275, -0.04010133817791939, 0.06601472944021225, -0.06500159949064255, 0.12984634935855865, -0.002928262809291482, 0.08684844523668289, 0.037593696266412735, -0.03463571146130562, -0.029077816754579544, -0.011839518323540688, 0.15655682981014252, 0.04033505916595459, -0.03314552828669548, 0.1066771149635315, 0.022321339696645737, 0.07313229888677597, 0.07412467896938324, 0.18051885068416595, 0.12964023649692535, 0.02457394078373909, 0.07551728934049606, 0.0456620417535305, -0.017998360097408295, -0.190249964594841, -0.0008996282122097909, 0.011068318970501423, 0.08670291304588318, -0.018140830099582672, 0.16476355493068695, 0.1411169469356537, -0.11325958371162415, 0.05500602349638939, -0.05284694582223892, -0.09324934333562851, -0.11732088029384613, -0.08692699670791626, -0.07868491858243942, -0.14460940659046173, 0.03139294311404228, -0.12993858754634857, 0.017682626843452454, 0.05836855247616768, 0.007623116951435804, -0.02115258015692234, 0.16096094250679016, -0.006314350292086601, 0.02816413715481758, 0.05296750366687775, -0.0035686951596289873, -0.02516467683017254, -0.029122134670615196, -0.03152279183268547, 0.02700585126876831, -0.01672952063381672, 0.07642584294080734, -0.01882617361843586, -0.018454713746905327, 0.027299746870994568, -0.035186767578125, -0.09861443191766739, 0.016008803620934486, 0.04360685124993324, 0.05960601940751076, 0.025381004437804222, 0.015107240527868271, -0.008917951956391335, -0.05354883894324303, 0.22268947958946228, -0.0736236721277237, -0.055910851806402206, -0.12049099057912827, 0.2747057378292084, 0.04235919564962387, -0.029555195942521095, 0.051600709557533264, -0.08649840205907822, -0.009770963340997696, 0.21303987503051758, 0.1540246456861496, -0.019023973494768143, -0.013484440743923187, -0.03297420218586922, -0.016873406246304512, -0.04655580222606659, 0.07991760224103928, 0.11517295986413956, 0.0072324080392718315, -0.08052454888820648, -0.017609750851988792, -0.06196044385433197, -0.016193101182579994, -0.08775696158409119, 0.045536383986473083, 0.039532121270895004, 0.008711477741599083, -0.058427274227142334, 0.03997275233268738, -0.06019727885723114, -0.11466830968856812, 0.04210728406906128, -0.16950711607933044, -0.1493833363056183, -0.004688954912126064, 0.044686734676361084, 0.00894483644515276, 0.09249349683523178, -0.010366630740463734, -0.008013752289116383, 0.09949804097414017, 0.0019196603680029511, -0.08430183678865433, -0.079144187271595, 0.13443970680236816, -0.11401502788066864, 0.18894267082214355, -0.012130978517234325, 0.08010172098875046, 0.13281486928462982, 0.021251238882541656, -0.1363552361726761, 0.02871292643249035, 0.046070680022239685, -0.032760538160800934, -0.010839683935046196, 0.15400122106075287, -0.009587944485247135, 0.06288863718509674, 0.02498778887093067, -0.15961027145385742, -0.03782147914171219, -0.040751080960035324, 0.02230549417436123, -0.05788690969347954, -0.014062249101698399, -0.0543336346745491, 0.13924163579940796, 0.1983790099620819, -0.050151508301496506, -0.0041871024295687675, -0.06961102038621902, 0.03160468861460686, 0.0765652284026146, -0.009191488847136497, -0.05818875879049301, -0.22748294472694397, 0.003899114904925227, 0.060273680835962296, -0.007665729615837336, -0.246900737285614, -0.08546900749206543, 0.022462986409664154, -0.054719697684049606, -0.07489357143640518, 0.06981998682022095, 0.057576414197683334, 0.0446808859705925, -0.04480810835957527, -0.049780119210481644, -0.08094865083694458, 0.14950354397296906, -0.15764527022838593, -0.041306398808956146 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # test-squad-trained-finetuned-squad This model was trained from scratch on the squad dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cu110 - Datasets 1.13.3 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "datasets": ["squad"], "model-index": [{"name": "test-squad-trained-finetuned-squad", "results": []}]}
question-answering
AyushPJ/test-squad-trained-finetuned-squad
[ "transformers", "pytorch", "tensorboard", "distilbert", "question-answering", "generated_from_trainer", "dataset:squad", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #question-answering #generated_from_trainer #dataset-squad #endpoints_compatible #region-us
# test-squad-trained-finetuned-squad This model was trained from scratch on the squad dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Framework versions - Transformers 4.11.3 - Pytorch 1.7.1+cu110 - Datasets 1.13.3 - Tokenizers 0.10.3
[ "# test-squad-trained-finetuned-squad\n\nThis model was trained from scratch on the squad dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cu110\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #question-answering #generated_from_trainer #dataset-squad #endpoints_compatible #region-us \n", "# test-squad-trained-finetuned-squad\n\nThis model was trained from scratch on the squad dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cu110\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ 48, 29, 6, 12, 8, 3, 90, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #question-answering #generated_from_trainer #dataset-squad #endpoints_compatible #region-us \n# test-squad-trained-finetuned-squad\n\nThis model was trained from scratch on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.7.1+cu110\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ -0.11998402327299118, 0.13743796944618225, -0.0014860376249998808, 0.09156715869903564, 0.13942953944206238, 0.02398286946117878, 0.14156971871852875, 0.11936116963624954, -0.08051753044128418, 0.07522286474704742, 0.11054574698209763, 0.07448318600654602, 0.03701362758874893, 0.141563281416893, -0.028644144535064697, -0.20955882966518402, 0.026677105575799942, 0.041556552052497864, -0.05014018714427948, 0.12027783691883087, 0.0940287858247757, -0.12900102138519287, 0.06536244601011276, 0.0008599695865996182, -0.2156524658203125, -0.009604659862816334, -0.01933206059038639, -0.05281951650977135, 0.10182810574769974, -0.010332678444683552, 0.13005869090557098, 0.0150993000715971, 0.08665987849235535, -0.15860562026500702, 0.01201449055224657, 0.05075510963797569, 0.026801927015185356, 0.09630138427019119, 0.025129301473498344, 0.03031964600086212, 0.041840020567178726, -0.11967378854751587, 0.0857260599732399, 0.016847865656018257, -0.09586373716592789, -0.14638929069042206, -0.10372506827116013, 0.06159340590238571, 0.06450791656970978, 0.09419465810060501, -0.010550612583756447, 0.20508086681365967, -0.07788094133138657, 0.07872550189495087, 0.19627131521701813, -0.3060217797756195, -0.0789007693529129, 0.06142780929803848, 0.07681813836097717, 0.045040663331747055, -0.10496508330106735, -0.02988317422568798, 0.045093026012182236, 0.06683392077684402, 0.0849984884262085, -0.027438675984740257, -0.08525358885526657, -0.009543509222567081, -0.15750285983085632, -0.03525460138916969, 0.16246047616004944, 0.00935019738972187, -0.04988699406385422, -0.10296062380075455, -0.07865353673696518, -0.09718252718448639, -0.02224338799715042, -0.054806679487228394, 0.012927509844303131, -0.060708753764629364, -0.10473017394542694, -0.03819115087389946, -0.08005847781896591, -0.04962712153792381, -0.04139385744929314, 0.07280468940734863, 0.046700362116098404, 0.033308591693639755, -0.06480731815099716, 0.09433916211128235, -0.01074459869414568, -0.12644164264202118, -0.008408412337303162, -0.019165804609656334, -0.06736022979021072, -0.06796403974294662, -0.03345951810479164, -0.07679715752601624, 0.01929665356874466, 0.15546096861362457, -0.06213778629899025, 0.040864963084459305, 0.017879923805594444, 0.005348088685423136, 0.0007385380449704826, 0.16069315373897552, -0.06703532487154007, -0.012420427054166794, 0.0016601344104856253, 0.08728030323982239, 0.013088724575936794, -0.013329808600246906, -0.09493324160575867, -0.025918399915099144, 0.12744371592998505, 0.06143099069595337, -0.012794328853487968, 0.03471333533525467, -0.014352618716657162, -0.04459410905838013, 0.02148900181055069, -0.10941300541162491, 0.04055438190698624, -0.026447422802448273, -0.0712612196803093, 0.014895009808242321, -0.01966317743062973, -0.015125690959393978, -0.04931033030152321, 0.03926176577806473, -0.10883443057537079, 0.014843062497675419, -0.07716186344623566, -0.09077488631010056, 0.03975934907793999, -0.1160004734992981, -0.024294544011354446, -0.0773942768573761, -0.15841859579086304, -0.015158485621213913, 0.025000831112265587, -0.0562724806368351, -0.04075081646442413, -0.03032364509999752, -0.07489745318889618, 0.01324418280273676, -0.0079220375046134, 0.0736471489071846, -0.029390400275588036, 0.08155159652233124, 0.04019898176193237, 0.04868751019239426, -0.013808838091790676, 0.03822999447584152, -0.10331359505653381, 0.04694695025682449, -0.17561841011047363, 0.0690142810344696, -0.07615280151367188, 0.036091066896915436, -0.12065327912569046, -0.1102176085114479, 0.05657678842544556, -0.04123019054532051, 0.09892784804105759, 0.13506363332271576, -0.16940344870090485, -0.016548318788409233, 0.16093863546848297, -0.07903904467821121, -0.11105526983737946, 0.12419503182172775, -0.05217715725302696, 0.02237313613295555, 0.057761840522289276, 0.16376301646232605, 0.09146679192781448, -0.12817876040935516, -0.05507578328251839, 0.008583693765103817, 0.042686063796281815, -0.05748378857970238, 0.0676681324839592, 0.021813344210386276, 0.04167093336582184, 0.015113976784050465, -0.028664806857705116, 0.01095871813595295, -0.10180103778839111, -0.08162245899438858, -0.06321823596954346, -0.08909223228693008, 0.08485233038663864, 0.04327360540628433, 0.0448165237903595, -0.0677391067147255, -0.0884297713637352, 0.10706190019845963, 0.10559339076280594, -0.059830911457538605, 0.014337276108562946, -0.10186604410409927, 0.116774782538414, -0.12614941596984863, -0.024367382749915123, -0.21671749651432037, -0.147049680352211, 0.04614404961466789, -0.011537169106304646, 0.028567204251885414, 0.03231958672404289, 0.05639924481511116, 0.06308945268392563, -0.03349308669567108, -0.016552025452256203, -0.06351817399263382, -0.017348911613225937, -0.13244739174842834, -0.14469899237155914, -0.07232213765382767, -0.04444606602191925, 0.12653328478336334, -0.19343696534633636, 0.018769092857837677, -0.0002449144667480141, 0.127089723944664, 0.007504672277718782, -0.03175275772809982, 0.006567773409187794, 0.055217940360307693, -0.021291950717568398, -0.08739232271909714, 0.05690005421638489, 0.009526357986032963, -0.072754867374897, -0.07553324848413467, -0.11846412718296051, 0.056019604206085205, 0.09291484951972961, 0.025304965674877167, -0.07885640114545822, -0.016608187928795815, -0.08066582679748535, -0.02912179008126259, -0.0758068636059761, -0.0035000196658074856, 0.13649816811084747, 0.02333941124379635, 0.12684319913387299, -0.07860366255044937, -0.06505688279867172, 0.03151007369160652, -0.0019983055535703897, -0.021611154079437256, 0.0793277770280838, 0.05676084756851196, -0.10407701879739761, 0.12533389031887054, 0.10927975922822952, -0.03891875594854355, 0.11331740021705627, -0.05542928725481033, -0.07981489598751068, -0.051591403782367706, -0.002837291220203042, -0.007662359159439802, 0.1416996866464615, -0.11440283805131912, -0.010439885780215263, 0.04493175819516182, 0.012344028800725937, 0.024599667638540268, -0.17548923194408417, -0.018691807985305786, 0.004058996681123972, -0.024203060194849968, -0.023573780432343483, 0.006810308899730444, 0.03530688211321831, 0.09713710844516754, 0.021353086456656456, -0.03991805389523506, 0.047344598919153214, 0.00021613429998978972, -0.0761093720793724, 0.1962490975856781, -0.09295491874217987, -0.1685047298669815, -0.10821836441755295, 0.0048796385526657104, -0.0723719671368599, -0.023467911407351494, 0.02069929987192154, -0.07826065272092819, -0.02627662755548954, -0.07676058262586594, -0.017483917996287346, -0.01989474892616272, 0.011843389831483364, 0.033852528780698776, -0.012170505709946156, 0.06821250170469284, -0.12279883772134781, 0.007767150644212961, -0.028956983238458633, -0.053030382841825485, 0.017588654533028603, 0.04062233120203018, 0.11470921337604523, 0.13813601434230804, -0.04129864647984505, 0.03152480721473694, -0.03678234666585922, 0.2849889099597931, -0.06966368854045868, -0.01137361116707325, 0.12746798992156982, 0.0005349111161194742, 0.05989463999867439, 0.1199335902929306, 0.037590328603982925, -0.08822567015886307, 0.013535158708691597, 0.025960994884371758, -0.03155193850398064, -0.2553066313266754, -0.039858151227235794, -0.04283732920885086, -0.016024522483348846, 0.09366722404956818, 0.029244186356663704, 0.018741048872470856, 0.08223749697208405, 0.005514673423022032, 0.06668420135974884, -0.06372764706611633, 0.08703697472810745, 0.12044136226177216, 0.035927072167396545, 0.10544285923242569, -0.04599374905228615, -0.07238325476646423, 0.037513770163059235, 0.018483109772205353, 0.25481709837913513, -0.016125988215208054, 0.1276269406080246, 0.03638017550110817, 0.17319560050964355, 0.0002834530605468899, 0.06963276863098145, -0.007678060792386532, -0.00513481767848134, 0.01570160500705242, -0.04999404028058052, -0.03471359610557556, -0.011166363023221493, 0.008342092856764793, 0.05744774639606476, -0.10210253298282623, 0.03986957669258118, 0.02659820020198822, 0.25922736525535583, 0.06331315636634827, -0.31923961639404297, -0.0999101921916008, -0.0037154685705900192, -0.030415652319788933, -0.06312131136655807, 0.005083966534584761, 0.14246773719787598, -0.12020894140005112, 0.020391272380948067, -0.09099139273166656, 0.08824039250612259, -0.030522126704454422, 0.006572810932993889, 0.04862162843346596, 0.11670845001935959, 0.005603345111012459, 0.08916802704334259, -0.22472895681858063, 0.2084427922964096, 0.030599728226661682, 0.08528319001197815, -0.09157513082027435, 0.004844387527555227, -0.0174629595130682, 0.054544512182474136, 0.1144910603761673, -0.0022588567808270454, -0.050946060568094254, -0.16502287983894348, -0.09914623945951462, 0.024900691583752632, 0.09489300847053528, 0.014624450355768204, 0.11552906781435013, -0.03115895576775074, 0.00600381288677454, 0.031078942120075226, -0.025691846385598183, -0.10059425234794617, -0.11801283061504364, 0.031661368906497955, 0.023791557177901268, -0.0749262273311615, -0.06218689680099487, -0.11344584822654724, -0.03785114735364914, 0.14848560094833374, -0.003333027707412839, -0.0524483323097229, -0.13151688873767853, 0.09171593934297562, 0.1398857980966568, -0.07143732905387878, 0.009526044130325317, -0.005731828976422548, 0.13389015197753906, 0.0008853389881551266, -0.09180253744125366, 0.051710739731788635, -0.06370215117931366, -0.1537882685661316, -0.044144876301288605, 0.12705841660499573, 0.04974403232336044, 0.058413948863744736, 0.015496594831347466, 0.0238573607057333, -0.002404494909569621, -0.08388987928628922, 0.008498994633555412, 0.022047419100999832, 0.08409064263105392, 0.03327331319451332, -0.037779323756694794, 0.04169870913028717, -0.04152128845453262, 0.02028438076376915, 0.1785135269165039, 0.17249411344528198, -0.09229802340269089, 0.050485990941524506, 0.07820356637239456, -0.06308817118406296, -0.19355084002017975, 0.06319920718669891, 0.07516244053840637, 0.020715493708848953, 0.011437677778303623, -0.15366999804973602, 0.11670520156621933, 0.08841975033283234, -0.007551556918770075, 0.0688752606511116, -0.3380300998687744, -0.10997685045003891, 0.080097995698452, 0.10566787421703339, 0.06168961897492409, -0.13675056397914886, -0.03527111932635307, -0.004229425452649593, -0.12674900889396667, 0.13749347627162933, -0.11610916256904602, 0.11168161034584045, -0.02630213089287281, 0.1236066147685051, 0.021224020048975945, -0.07045994699001312, 0.1239265576004982, 0.0560247041285038, 0.07789897173643112, -0.055666111409664154, -0.03714670240879059, 0.16554966568946838, -0.06541252881288528, 0.07377958297729492, 0.004258353263139725, 0.05881551653146744, -0.1397741734981537, -0.018763253465294838, -0.08048742264509201, 0.04390599951148033, -0.05170399695634842, -0.051214203238487244, -0.050046734511852264, 0.06051187589764595, 0.04480946809053421, -0.030505837872624397, 0.09219620376825333, 0.05803731456398964, 0.13135123252868652, 0.10155902802944183, 0.11530567705631256, -0.03146634250879288, -0.11342156678438187, 0.009283656254410744, -0.0175669863820076, 0.05233640596270561, -0.08891031891107559, 0.011843916960060596, 0.16047777235507965, 0.04644254222512245, 0.11845673620700836, 0.05753868445754051, -0.06149357184767723, 0.007818876765668392, 0.02096121571958065, -0.14615844190120697, -0.18410661816596985, 0.0007400280446745455, -0.010304947383701801, -0.17673076689243317, 0.05465884879231453, 0.10513443499803543, -0.06146467849612236, -0.03397207334637642, -0.007583441212773323, -0.007054466288536787, -0.01654699072241783, 0.19950072467327118, 0.06284700334072113, 0.07771138846874237, -0.09994816035032272, 0.08446282893419266, 0.07253477722406387, -0.0992252379655838, 0.042916033416986465, 0.06989304721355438, -0.09546816349029541, -0.032255869358778, 0.057124122977256775, 0.1533975601196289, -0.07603563368320465, -0.02566065825521946, -0.1039140522480011, -0.09094689041376114, 0.0563313290476799, 0.16566729545593262, 0.07097718119621277, 0.004524468909949064, -0.04326820746064186, 0.01079296600073576, -0.15938550233840942, 0.10822468250989914, 0.037412796169519424, 0.07691936194896698, -0.138152614235878, 0.11720731109380722, -0.01289567444473505, 0.09233244508504868, -0.03359387814998627, 0.011422713287174702, -0.11093151569366455, -0.00513844471424818, -0.14619819819927216, -0.0031271292828023434, -0.054551202803850174, -0.002943150233477354, -0.022676929831504822, -0.04710347205400467, -0.0407683365046978, 0.029442433267831802, -0.0703200176358223, -0.03330295905470848, 0.025739016011357307, 0.01687740907073021, -0.14557531476020813, -0.004987239371985197, -0.0003929889062419534, -0.07660596072673798, 0.09253738820552826, 0.05254011228680611, 0.008609761483967304, 0.015352622605860233, -0.037710390985012054, -0.03948471322655678, 0.0015035372925922275, 0.025832122191786766, 0.10303949564695358, -0.0875432938337326, -0.004719902295619249, -0.005053216125816107, 0.041883256286382675, 0.04426930472254753, 0.05562400445342064, -0.11750822514295578, -0.02473226934671402, -0.04212144762277603, -0.03838113695383072, -0.07573588937520981, 0.05410002917051315, 0.10199252516031265, 0.038086339831352234, 0.16507010161876678, -0.07130905985832214, 0.0686885416507721, -0.19514650106430054, -0.032952483743429184, -0.007865071296691895, -0.05848109349608421, -0.03698517382144928, -0.03374747559428215, 0.07399438321590424, -0.0660281777381897, 0.11766356974840164, -0.021197501569986343, 0.108549103140831, 0.03120461106300354, 0.002802149625495076, -0.024136831983923912, 0.010561543516814709, 0.15879353880882263, 0.03564660623669624, -0.03715105727314949, 0.08227560669183731, 0.018664004281163216, 0.06242511793971062, 0.10823813825845718, 0.1906324028968811, 0.1202494204044342, 0.018227381631731987, 0.07645516842603683, 0.07464241236448288, -0.04341525211930275, -0.14583218097686768, 0.04930247738957405, -0.03127454221248627, 0.09463384002447128, -0.03240426629781723, 0.17037764191627502, 0.08933735638856888, -0.14868994057178497, 0.05297970771789551, -0.0663362368941307, -0.12142446637153625, -0.09557320922613144, -0.0853293314576149, -0.07015170156955719, -0.09388456493616104, 0.011459018103778362, -0.1323917806148529, -0.006317250896245241, 0.09658718854188919, 0.024189848452806473, -0.02249910682439804, 0.17933861911296844, 0.015706894919276237, 0.03566157817840576, 0.067535400390625, 0.021408384665846825, 0.010364094749093056, -0.030055569484829903, -0.021643593907356262, 0.027345571666955948, 0.012868564575910568, 0.0842038169503212, -0.030893994495272636, 0.01509302482008934, 0.03621581196784973, -0.020663609728217125, -0.09085895121097565, 0.006111782044172287, 0.03549383580684662, 0.04478690028190613, 0.020386697724461555, 0.033825233578681946, 0.006092668976634741, -0.05247439816594124, 0.19626837968826294, -0.05952021852135658, -0.0914982408285141, -0.1399495005607605, 0.20345576107501984, 0.001548828324303031, -0.03297591209411621, 0.06515491008758545, -0.1022672951221466, -0.011432824656367302, 0.1933862268924713, 0.17672188580036163, -0.06813260167837143, -0.021835310384631157, 0.00839556846767664, -0.01707799918949604, -0.0501076877117157, 0.11169462651014328, 0.0995001494884491, 0.07241913676261902, -0.08257844299077988, -0.027630649507045746, -0.05156364664435387, -0.04216868057847023, -0.061473701149225235, 0.04484076052904129, 0.04005226492881775, 0.029929298907518387, -0.04792117327451706, 0.06619124859571457, -0.0632229596376419, -0.10578979551792145, 0.0355755090713501, -0.16250498592853546, -0.17972305417060852, -0.01602347567677498, 0.06590291857719421, -0.01864655129611492, 0.05906401574611664, -0.026137148961424828, 0.018693266436457634, 0.12152939289808273, -0.01896193064749241, -0.06755328178405762, -0.07157508283853531, 0.1283571720123291, -0.07639963179826736, 0.17622868716716766, -0.014619109220802784, 0.10592735558748245, 0.1348544955253601, 0.015133854001760483, -0.14685988426208496, 0.006801656913012266, 0.08421836048364639, -0.04182446748018265, 0.013937321491539478, 0.14383326470851898, -0.006686699111014605, 0.06986720114946365, 0.03796055540442467, -0.12775911390781403, -0.026249343529343605, -0.04722938686609268, 0.0250223521143198, -0.09202715754508972, 0.0064932298846542835, -0.052942804992198944, 0.1556129902601242, 0.20028617978096008, -0.05154348909854889, -0.003134919796139002, -0.06587646156549454, 0.005104687064886093, 0.04670261964201927, 0.0541277639567852, -0.04174911603331566, -0.23125943541526794, 0.0171770378947258, 0.010338430292904377, 0.010380827821791172, -0.2458491176366806, -0.08725114166736603, 0.04870468005537987, -0.05186885967850685, -0.07358122617006302, 0.10146655142307281, 0.06163116544485092, 0.05436006188392639, -0.04778950288891792, -0.04308651387691498, -0.08514341711997986, 0.14222528040409088, -0.15341617166996002, -0.056589435786008835 ]
null
null
transformers
# Echidona DialoGPT-Medium Model
{"tags": ["conversational"]}
text-generation
Azuris/DialoGPT-medium-envy
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Echidona DialoGPT-Medium Model
[ "# Echidona DialoGPT-Medium Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Echidona DialoGPT-Medium Model" ]
[ 51, 12 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Echidona DialoGPT-Medium Model" ]
[ -0.03381689265370369, -0.004181369207799435, -0.007092528976500034, 0.019710838794708252, 0.13752590119838715, 0.009560720063745975, 0.13168564438819885, 0.10607534646987915, -0.005329482723027468, -0.016730966046452522, 0.1059640571475029, 0.17826882004737854, -0.004718042444437742, 0.11881506443023682, -0.07289060950279236, -0.3134220540523529, 0.035145971924066544, 0.06694246083498001, 0.020464638248085976, 0.11579351872205734, 0.11519799381494522, -0.05403755232691765, 0.08465220034122467, 0.0063630067743361, -0.1282106637954712, 0.02600228227674961, -0.0033838292583823204, -0.1250542402267456, 0.11102273315191269, 0.0701756700873375, 0.03512889891862869, 0.016095571219921112, -0.034607771784067154, -0.12896351516246796, 0.028708573430776596, -0.030350692570209503, -0.0480072908103466, 0.02773660607635975, -0.0002013694029301405, -0.09269518405199051, 0.13961158692836761, 0.06861045211553574, 0.016050036996603012, 0.004563902039080858, -0.14140044152736664, 0.03000321239233017, 0.015284465625882149, 0.013180241920053959, 0.07974113523960114, 0.10476861894130707, -0.04070035368204117, 0.12029363214969635, -0.06887954473495483, 0.10051914304494858, 0.09124673902988434, -0.3298237919807434, -0.021535681560635567, 0.06719058007001877, 0.05560560151934624, 0.08891793340444565, -0.02183382771909237, 0.08032044023275375, 0.04139731079339981, 0.02360169216990471, -0.011242230422794819, -0.06173236295580864, -0.06751615554094315, -0.00566300330683589, -0.09656795114278793, -0.004961324390023947, 0.25545674562454224, -0.04671049863100052, 0.05398442968726158, -0.0753224715590477, -0.09813486784696579, -0.013868704438209534, -0.033214863389730453, -0.04674116522073746, -0.06608118861913681, 0.060899216681718826, 0.015809010714292526, -0.08117581158876419, -0.13044121861457825, -0.015773693099617958, -0.1851191520690918, 0.19544504582881927, 0.040052905678749084, 0.027849450707435608, -0.20151972770690918, 0.1019420474767685, 0.01796574890613556, -0.0955834612250328, 0.016342364251613617, -0.07658394426107407, 0.037405867129564285, 0.007360886782407761, -0.021204130724072456, -0.057387370616197586, 0.05125449225306511, 0.1429458111524582, -0.022446412593126297, 0.010164284147322178, 0.0016361786983907223, 0.04282842203974724, 0.05527748912572861, 0.05468885600566864, -0.003270471701398492, -0.05995230749249458, 0.019662562757730484, -0.10150299966335297, 0.025383392348885536, -0.05784044414758682, -0.17780464887619019, -0.029056783765554428, 0.022080576047301292, 0.05408510938286781, 0.011409829370677471, 0.10045695304870605, -0.015159722417593002, -0.04856439307332039, 0.0614335797727108, -0.033247027546167374, -0.002790342550724745, 0.00015128993254620582, -0.0024316515773534775, 0.16838914155960083, 0.011251943185925484, 0.04094674438238144, -0.12384101748466492, 0.025787929072976112, -0.053995054215192795, -0.02141132764518261, -0.02964199148118496, -0.05171250179409981, 0.01015591248869896, -0.026125112548470497, 0.026118261739611626, -0.16754013299942017, -0.14148738980293274, 0.002722963457927108, -0.011945489794015884, -0.06842004507780075, -0.09900941699743271, -0.1045318990945816, 0.010021737776696682, 0.06362444162368774, -0.07916311919689178, -0.011847937479615211, -0.05934397131204605, 0.07168255001306534, -0.0011995116947218776, 0.10077285766601562, -0.09735304862260818, 0.06292810291051865, -0.06254840642213821, -0.03843363747000694, -0.07784681022167206, 0.09062203019857407, 0.02098606899380684, 0.05720075964927673, -0.012388193979859352, -0.020259911194443703, -0.09659870713949203, 0.08178494870662689, -0.005626197904348373, 0.22604992985725403, -0.05692170187830925, -0.1102202981710434, 0.2632279694080353, -0.05665258690714836, -0.12123608589172363, 0.12544924020767212, 0.014225328341126442, 0.07891348004341125, 0.13760074973106384, 0.15186850726604462, -0.05161545053124428, -0.040728386491537094, 0.07191212475299835, 0.08315162360668182, -0.06582463532686234, -0.00959788728505373, 0.02025824598968029, -0.01756708323955536, -0.04974866285920143, 0.028476083651185036, 0.09724466502666473, 0.040576137602329254, -0.07217074185609818, -0.00802928488701582, 0.015971044078469276, 0.0015666938852518797, 0.07620898634195328, -0.05641238018870354, 0.11958380043506622, -0.0498746782541275, -0.06393998861312866, 0.003088026773184538, 0.021683836355805397, -0.061526063829660416, 0.05277661606669426, -0.07819484174251556, 0.08850862830877304, -0.028362326323986053, 0.057494230568408966, -0.12813253700733185, -0.046132802963256836, -0.02363833785057068, 0.17555029690265656, 0.05582704767584801, 0.06625111401081085, 0.03964119032025337, -0.04112422466278076, -0.019003773108124733, 0.03567707538604736, 0.13900695741176605, -0.020371930673718452, -0.08725219964981079, -0.10364752262830734, 0.10226308554410934, -0.05738399177789688, 0.07980944216251373, -0.07592253386974335, 0.00970273744314909, 0.04635864868760109, 0.10592097043991089, -0.020546438172459602, 0.054167766124010086, -0.010709825903177261, 0.0035390760749578476, -0.06419549882411957, 0.027050292119383812, 0.07692910730838776, -0.0027032142970710993, -0.09637537598609924, 0.20716434717178345, -0.17673669755458832, 0.1504775732755661, 0.1791049689054489, -0.24282971024513245, 0.014258837327361107, -0.1334688514471054, -0.016933126375079155, 0.00347807421348989, 0.04498782008886337, -0.05469360575079918, 0.20196925103664398, -0.03556303679943085, 0.17831307649612427, -0.0535678006708622, -0.022228267043828964, -0.024239404127001762, -0.06476720422506332, -0.0010243791621178389, 0.09517494589090347, 0.10383488237857819, -0.11254129558801651, 0.14007815718650818, 0.10173112154006958, 0.026905501261353493, 0.21222929656505585, 0.06322351098060608, -0.0028040127363055944, 0.05458837375044823, -0.019597912207245827, -0.044131625443696976, -0.07024184614419937, -0.2719901502132416, -0.03276009485125542, 0.06307678669691086, 0.052241306751966476, 0.12702767550945282, -0.08980976790189743, -0.015884600579738617, 0.013799921609461308, -0.014832108281552792, 0.058982670307159424, 0.11607632040977478, 0.004270175006240606, 0.11543162912130356, -0.02175968885421753, -0.045817822217941284, 0.05503268167376518, 0.022017939016222954, -0.09023651480674744, 0.19739298522472382, -0.1423410177230835, -0.33717307448387146, -0.10068652033805847, -0.1696992665529251, -0.0464550144970417, 0.060709960758686066, 0.11726353317499161, -0.1138763278722763, -0.02609981596469879, 0.006242986302822828, 0.09926316142082214, -0.05816483497619629, 0.016101054847240448, -0.006417241413146257, -0.00678686099126935, -0.13593317568302155, -0.09979782998561859, -0.0608818344771862, -0.043988876044750214, -0.06720520555973053, 0.1294340193271637, -0.14971740543842316, 0.061396919190883636, 0.22956803441047668, 0.060745567083358765, 0.040916044265031815, -0.032835204154253006, 0.19784849882125854, -0.12227965146303177, 0.0032653617672622204, 0.1801312267780304, -0.017749210819602013, 0.04966295510530472, 0.12134718894958496, -0.0030173007398843765, -0.08235277235507965, 0.03682868555188179, -0.03104293905198574, -0.06543342024087906, -0.2291208952665329, -0.14900198578834534, -0.11016138643026352, 0.041791804134845734, 0.01439063623547554, 0.04166639596223831, 0.12455658614635468, 0.06220537796616554, -0.03482889011502266, 0.012009022757411003, 0.06166280061006546, 0.08955778181552887, 0.26069650053977966, -0.06553584337234497, 0.13894008100032806, -0.0047718859277665615, -0.16193312406539917, 0.09053150564432144, 0.10336355865001678, 0.0991184338927269, 0.07525470107793808, 0.08701156079769135, 0.03363104537129402, -0.020901711657643318, 0.11644540727138519, 0.07172857969999313, 0.010139438323676586, -0.03355822712182999, -0.0421845018863678, -0.046708520501852036, -0.002726671053096652, 0.042603619396686554, 0.038276005536317825, -0.13534247875213623, -0.07222513854503632, -0.019683988764882088, 0.08014741539955139, 0.025931091979146004, 0.07710788398981094, -0.18612828850746155, -0.009952335618436337, 0.07126050442457199, -0.02662229910492897, -0.11286603659391403, 0.07333601266145706, 0.006775776389986277, -0.12558814883232117, 0.06673852354288101, 0.0048818811774253845, 0.10984333604574203, -0.09049031138420105, 0.07143816351890564, -0.10114041715860367, -0.058586400002241135, -0.007020933553576469, 0.0882035344839096, -0.2735820710659027, 0.21839062869548798, -0.010163825936615467, -0.03629916161298752, -0.09830076992511749, -0.007999961264431477, -0.002897468861192465, 0.12812864780426025, 0.09192265570163727, -0.02381448820233345, 0.0383879691362381, 0.01589849591255188, -0.05186512693762779, 0.038927819579839706, 0.09333311766386032, -0.02476259507238865, 0.0015054511604830623, -0.006856204010546207, 0.00007648332393728197, -0.012287982739508152, -0.0749620720744133, -0.042163196951150894, -0.1938266009092331, 0.07578730583190918, 0.08573923259973526, 0.05817297846078873, 0.033536944538354874, -0.05663933604955673, -0.06588636338710785, 0.21866799890995026, -0.03159533813595772, -0.10386152565479279, -0.08318892121315002, -0.02707398496568203, 0.04149249941110611, -0.06511414796113968, 0.03550076112151146, -0.03149663284420967, 0.05388068035244942, -0.06355863064527512, -0.14447975158691406, 0.11797437071800232, -0.10681723058223724, -0.03898348659276962, -0.03209353983402252, 0.19448038935661316, 0.004560387693345547, 0.004128197208046913, 0.054115988314151764, -0.02661719173192978, -0.10598912090063095, -0.08732501417398453, -0.023283887654542923, 0.054133761674165726, 0.02163899503648281, 0.047360338270664215, -0.054217517375946045, -0.07666272670030594, -0.05371928587555885, -0.048727940768003464, 0.28801247477531433, 0.13714414834976196, -0.018439605832099915, 0.14587432146072388, 0.17322030663490295, -0.08703997731208801, -0.24456968903541565, -0.1228383332490921, -0.05281335860490799, -0.004380315076559782, -0.07833259552717209, -0.1647920310497284, 0.06552667915821075, -0.01187429204583168, -0.01526406779885292, 0.09203842282295227, -0.29598650336265564, -0.10503567010164261, 0.14683866500854492, -0.030437523499131203, 0.37813857197761536, -0.08981504291296005, -0.08216172456741333, -0.07054296135902405, -0.18087080121040344, 0.14295069873332977, -0.01292748935520649, 0.10799527168273926, -0.0173389483243227, 0.17140741646289825, 0.03765488788485527, 0.020179836079478264, 0.0994080901145935, 0.040680695325136185, -0.06140805035829544, -0.08127286285161972, -0.043847087770700455, -0.027931492775678635, 0.025240981951355934, 0.04337959736585617, -0.06353961676359177, -0.0043502929620444775, -0.15120458602905273, -0.0815633237361908, -0.05994626507163048, 0.03323772922158241, 0.019085604697465897, -0.06599726527929306, 0.0005255513824522495, -0.07866298407316208, -0.00314560579136014, 0.01054172683507204, 0.12228231132030487, -0.0957745686173439, 0.11892496049404144, 0.06630357354879379, 0.14166295528411865, -0.20444291830062866, -0.011410201899707317, -0.05443393439054489, -0.057793907821178436, 0.07718028873205185, -0.055912427604198456, 0.021243570372462273, 0.1206824779510498, -0.055189188569784164, 0.08032499253749847, 0.0977141410112381, -0.014519047923386097, 0.013884393498301506, 0.11053645610809326, -0.2158256471157074, 0.007353333756327629, -0.05862809345126152, 0.022583043202757835, 0.0970265194773674, 0.10253532975912094, 0.21716144680976868, 0.0050123087130486965, -0.04306454584002495, -0.005383146461099386, 0.026208167895674706, -0.03504655882716179, 0.10046231746673584, 0.002297131810337305, -0.0012874627718701959, -0.12918192148208618, 0.046500496566295624, 0.012411529198288918, -0.12902356684207916, 0.024044228717684746, 0.1515851467847824, -0.09074386954307556, -0.11742501705884933, -0.052349869161844254, 0.09085231274366379, -0.16191540658473969, -0.028808370232582092, -0.03627606853842735, -0.14629749953746796, 0.070513054728508, 0.06665787845849991, 0.062186818569898605, 0.04743364453315735, -0.09025994688272476, -0.0011927509913221002, -0.014530407264828682, 0.0028853043913841248, 0.06676243245601654, -0.00993378460407257, -0.04950810968875885, 0.04848926514387131, -0.027329618111252785, 0.10115162283182144, -0.08609183132648468, -0.08046933263540268, -0.14396384358406067, 0.042871713638305664, -0.07894507050514221, -0.07114335894584656, -0.0791727602481842, -0.05977524444460869, -0.0038834228180348873, -0.016307547688484192, -0.052782416343688965, -0.037717048078775406, -0.09611336141824722, 0.03984758257865906, -0.03724095970392227, 0.01958547905087471, -0.05765139311552048, 0.013988745398819447, 0.043916426599025726, -0.038094233721494675, 0.15424539148807526, 0.12740106880664825, -0.10040713101625443, 0.07416345179080963, -0.16890937089920044, -0.045990802347660065, 0.11065608263015747, 0.03307134658098221, 0.060925375670194626, 0.062277380377054214, 0.01201725099235773, 0.06933697313070297, 0.0629347637295723, 0.04451821371912956, 0.01437039952725172, -0.08435395359992981, 0.03858288750052452, -0.05524984002113342, -0.1467037796974182, -0.03309008479118347, -0.014166546985507011, 0.009883849881589413, 0.02954334020614624, 0.07309625297784805, -0.06742634624242783, 0.05825057998299599, -0.04692063853144646, 0.046412937343120575, 0.019071320071816444, -0.14139017462730408, 0.004377895500510931, -0.07421889156103134, 0.04807371273636818, 0.007994323968887329, 0.20927490293979645, 0.032165855169296265, 0.0007794430712237954, 0.01950237713754177, 0.04028424248099327, 0.004108229652047157, 0.003987894859164953, 0.15935666859149933, 0.09830416738986969, -0.021699408069252968, -0.12089468538761139, 0.06731119006872177, 0.0632188618183136, 0.054789189249277115, 0.10484643280506134, -0.05465369671583176, 0.017934655770659447, 0.12288326770067215, 0.0008773374720476568, 0.02760588750243187, -0.16091711819171906, -0.18387079238891602, -0.03252735733985901, 0.058656252920627594, -0.03298087790608406, 0.0977279469370842, 0.1443881243467331, -0.02019805647432804, 0.0026519102975726128, -0.012405285611748695, -0.07085137814283371, -0.1833837926387787, -0.1759652942419052, -0.07930907607078552, -0.14158253371715546, -0.0030863306019455194, -0.1473371386528015, 0.03948906064033508, -0.0006304210983216763, 0.0892755463719368, -0.0642126277089119, 0.07610306143760681, 0.05183706060051918, -0.10972628742456436, 0.0735267773270607, -0.01967230997979641, 0.10528149455785751, -0.029880378395318985, 0.009422732517123222, -0.09462287276983261, 0.05754322558641434, -0.004327723290771246, 0.03831307962536812, -0.05497303977608681, 0.0060934810899198055, -0.08363205939531326, -0.05458185449242592, -0.0579889751970768, 0.054166410118341446, 0.010981175117194653, 0.14936845004558563, 0.005059692542999983, -0.0552985705435276, 0.028671951964497566, 0.21629826724529266, -0.06144609674811363, -0.09835603088140488, -0.0694383755326271, 0.17845454812049866, 0.031781330704689026, 0.133064404129982, -0.025440705940127373, 0.0021258932538330555, -0.11314651370048523, 0.31296831369400024, 0.2886866629123688, -0.09442799538373947, 0.006667249370366335, 0.020851531997323036, 0.04599279910326004, 0.1175575852394104, 0.08877094089984894, 0.11699704825878143, 0.3560023009777069, -0.03799852728843689, -0.015537945553660393, -0.030455878004431725, -0.0013092916924506426, -0.0782470554113388, 0.05451611429452896, 0.04818981513381004, -0.056943923234939575, -0.008365160785615444, 0.10893108695745468, -0.19272808730602264, 0.09112685918807983, -0.14840227365493774, -0.18658967316150665, -0.07834413647651672, 0.01750793494284153, 0.09509358555078506, 0.027611255645751953, 0.09706032276153564, -0.011744263581931591, -0.10296430438756943, 0.0932549387216568, 0.028577547520399094, -0.17648646235466003, 0.007400537375360727, 0.08790118247270584, -0.057031549513339996, -0.01167240459471941, -0.02663796953856945, 0.10984308272600174, 0.07202721387147903, 0.061100609600543976, 0.007259470410645008, 0.04101265221834183, -0.006469581741839647, -0.02881922572851181, 0.07248103618621826, 0.01816233992576599, 0.009369881823658943, -0.11129352450370789, 0.07537884265184402, -0.1513104885816574, 0.040932174772024155, 0.03320698440074921, -0.02382582053542137, -0.013996243476867676, 0.04370058327913284, -0.08764486759901047, 0.06254572421312332, 0.07224228978157043, -0.010971637442708015, -0.016882263123989105, -0.0032343664206564426, 0.011380844749510288, -0.016022583469748497, -0.024007366970181465, -0.0712726041674614, -0.1976851522922516, -0.09608156979084015, 0.06725065410137177, 0.01602628268301487, -0.1803426593542099, 0.0010218863608315587, -0.1320839673280716, 0.056032050400972366, -0.14783847332000732, 0.09128440916538239, 0.0758049488067627, 0.010276838205754757, -0.007769518997520208, 0.001051921397447586, 0.01788182370364666, 0.0787227675318718, -0.14219501614570618, -0.08557290583848953 ]
null
null
transformers
# Echidona DialoGPT Model
{"tags": ["conversational"]}
text-generation
Azuris/DialoGPT-small-envy
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Echidona DialoGPT Model
[ "# Echidona DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Echidona DialoGPT Model" ]
[ 51, 9 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Echidona DialoGPT Model" ]
[ -0.03286154940724373, 0.024418439716100693, -0.006946341600269079, 0.030053293332457542, 0.1391144096851349, 0.008320393040776253, 0.11900372803211212, 0.1091674342751503, 0.013224707916378975, -0.0091220922768116, 0.11974117159843445, 0.19499742984771729, -0.006253214552998543, 0.11450549960136414, -0.08339207619428635, -0.3045554757118225, 0.052807144820690155, 0.06427810341119766, 0.03261629119515419, 0.10843705385923386, 0.1071857213973999, -0.05431227385997772, 0.09184680879116058, -0.000190477236174047, -0.126573845744133, 0.028297938406467438, -0.004373075440526009, -0.12754766643047333, 0.1216387152671814, 0.07221242785453796, 0.04710226133465767, 0.011542542837560177, -0.035884782671928406, -0.1370261311531067, 0.02612961083650589, -0.02989964932203293, -0.056205350905656815, 0.03417898714542389, 0.005504818633198738, -0.09480676800012589, 0.12160184979438782, 0.08098121732473373, 0.014175176620483398, 0.007652781438082457, -0.15110068023204803, 0.04490978270769119, 0.018563609570264816, 0.000852431170642376, 0.08795721083879471, 0.10368240624666214, -0.04387960210442543, 0.11499731987714767, -0.08817598968744278, 0.09807472676038742, 0.09610296785831451, -0.32592031359672546, -0.015476624481379986, 0.061838798224925995, 0.05281304940581322, 0.06821756064891815, -0.03326787054538727, 0.07366370409727097, 0.04961800575256348, 0.014704600907862186, -0.014213193207979202, -0.05981309339404106, -0.09217338263988495, -0.0013691671192646027, -0.09678811579942703, -0.014273470267653465, 0.2540515065193176, -0.03739862143993378, 0.05623079463839531, -0.06667474657297134, -0.10861596465110779, -0.019084548577666283, -0.028630221262574196, -0.04692632704973221, -0.06658179312944412, 0.06276272982358932, 0.02565155178308487, -0.07490129768848419, -0.13054205477237701, -0.010398386046290398, -0.18554340302944183, 0.19602888822555542, 0.04451799392700195, 0.024424318224191666, -0.20487716794013977, 0.09807413071393967, 0.016236770898103714, -0.09447452425956726, 0.007148238364607096, -0.07354269921779633, 0.03196021914482117, 0.010046500712633133, -0.026492523029446602, -0.028495250269770622, 0.05597201734781265, 0.16046401858329773, -0.015293958596885204, 0.010093145072460175, 0.0005607351195067167, 0.04827523231506348, 0.0542820580303669, 0.07123520970344543, 0.005415511783212423, -0.05900416523218155, 0.019059736281633377, -0.10937576740980148, 0.018879882991313934, -0.06292298436164856, -0.18788914382457733, -0.0349758043885231, 0.032956257462501526, 0.062175314873456955, 0.011245743371546268, 0.10121994465589523, -0.023781057447195053, -0.040744613856077194, 0.05962429940700531, -0.03588170185685158, 0.0006888284115120769, 0.007388392463326454, -0.01023067906498909, 0.1428811252117157, 0.009256545454263687, 0.04404813051223755, -0.12487797439098358, 0.025903873145580292, -0.059995293617248535, -0.028223715722560883, -0.02764144167304039, -0.0475701168179512, 0.007502819877117872, -0.034334082156419754, 0.03182104974985123, -0.1650574803352356, -0.1470482349395752, 0.003415070939809084, -0.020153163000941277, -0.06877867877483368, -0.09626530110836029, -0.12356366217136383, 0.0067797498777508736, 0.05336734652519226, -0.07513341307640076, -0.022198744118213654, -0.0611235648393631, 0.07013044506311417, -0.008891476318240166, 0.10231645405292511, -0.10368471592664719, 0.06361458450555801, -0.08051948249340057, -0.03692363202571869, -0.05072702467441559, 0.08922126144170761, 0.029715068638324738, 0.0515659898519516, -0.011331292800605297, -0.018004707992076874, -0.09840178489685059, 0.08152969181537628, -0.012889515608549118, 0.23186135292053223, -0.057157792150974274, -0.10641781985759735, 0.27303269505500793, -0.053787216544151306, -0.1248212531208992, 0.12802541255950928, 0.006497091148048639, 0.07828052341938019, 0.13238680362701416, 0.16408313810825348, -0.07564512640237808, -0.030554641038179398, 0.07358235865831375, 0.08387106657028198, -0.07532042264938354, -0.003996404819190502, 0.018559809774160385, -0.024666188284754753, -0.04981957748532295, 0.029049066826701164, 0.0883428230881691, 0.03704812750220299, -0.0651555210351944, -0.015548283234238625, 0.013440809212625027, -0.000448581762611866, 0.07669614255428314, -0.05151724815368652, 0.1150139644742012, -0.04184548556804657, -0.06430276483297348, -0.011073168367147446, 0.02364542707800865, -0.06805820018053055, 0.04799608141183853, -0.07469503581523895, 0.09912405163049698, -0.023059148341417313, 0.05088826268911362, -0.12233807891607285, -0.06390029191970825, -0.012207236140966415, 0.15618585050106049, 0.04906424880027771, 0.06625571846961975, 0.040227022022008896, -0.02975338324904442, -0.01150607131421566, 0.024732418358325958, 0.14107875525951385, -0.022464871406555176, -0.0876910611987114, -0.10016623139381409, 0.09225993603467941, -0.05656474828720093, 0.09169028699398041, -0.09290184080600739, 0.004837842658162117, 0.03393714129924774, 0.11167345941066742, -0.020940497517585754, 0.05261920392513275, -0.01713799685239792, -0.005022753030061722, -0.07834438979625702, 0.02450624667108059, 0.0786832794547081, -0.010013100691139698, -0.08997313678264618, 0.19964107871055603, -0.17651206254959106, 0.13802365958690643, 0.1782689094543457, -0.24475568532943726, 0.011491295881569386, -0.13045354187488556, -0.023210398852825165, 0.0028497681487351656, 0.04409150406718254, -0.059641335159540176, 0.20388227701187134, -0.030615175142884254, 0.18386264145374298, -0.0556069016456604, -0.025845903903245926, -0.02500855177640915, -0.06171514093875885, -0.0017606018809601665, 0.09255243092775345, 0.12714682519435883, -0.1159360483288765, 0.15331940352916718, 0.09813684225082397, 0.0294942706823349, 0.20520322024822235, 0.06140691041946411, -0.01026059128344059, 0.05613071471452713, -0.007638959214091301, -0.045916855335235596, -0.06717050075531006, -0.27739638090133667, -0.029918236657977104, 0.06784754991531372, 0.04412613809108734, 0.12299913167953491, -0.09912891685962677, -0.019435100257396698, 0.009469123557209969, -0.016036994755268097, 0.0584111250936985, 0.10496832430362701, 0.007271969690918922, 0.11742782592773438, -0.021671529859304428, -0.07623987644910812, 0.06043887883424759, 0.02374737337231636, -0.08418770134449005, 0.20400643348693848, -0.13458625972270966, -0.3396998345851898, -0.10105939209461212, -0.17006276547908783, -0.04838430881500244, 0.06948354840278625, 0.11324954032897949, -0.11303866654634476, -0.028743138536810875, 0.010369493626058102, 0.1080050840973854, -0.062015026807785034, 0.011371592059731483, -0.026675531640648842, -0.00374526996165514, -0.148993119597435, -0.10394655168056488, -0.06092158704996109, -0.047751717269420624, -0.06507652997970581, 0.12603025138378143, -0.15220800042152405, 0.06108818203210831, 0.23499292135238647, 0.0572858490049839, 0.03853460028767586, -0.03207184374332428, 0.20336848497390747, -0.12034166604280472, 0.0028703552670776844, 0.15942542254924774, -0.0355733223259449, 0.053593941032886505, 0.12920276820659637, -0.00895974412560463, -0.09913130104541779, 0.039236485958099365, -0.0338737815618515, -0.06255978345870972, -0.23546409606933594, -0.14641830325126648, -0.10755323618650436, 0.05625113844871521, 0.032373301684856415, 0.04223024845123291, 0.15029582381248474, 0.07364465296268463, -0.040054645389318466, 0.01132466085255146, 0.04563518241047859, 0.09092066437005997, 0.24801741540431976, -0.06279470026493073, 0.13420109450817108, 0.0028336411342024803, -0.17332614958286285, 0.08148626983165741, 0.09529757499694824, 0.10102982819080353, 0.07743895053863525, 0.10158585757017136, 0.0211431086063385, 0.012750456109642982, 0.11572625488042831, 0.07684769481420517, 0.014446060173213482, -0.031971294432878494, -0.04539010673761368, -0.043546926230192184, -0.011598764918744564, 0.05853428691625595, 0.049821726977825165, -0.13982532918453217, -0.07290182262659073, -0.03603135049343109, 0.07582473009824753, 0.02425968647003174, 0.08800950646400452, -0.19904515147209167, -0.005888286978006363, 0.07336239516735077, -0.025978000834584236, -0.11886194348335266, 0.08017769455909729, -0.002616004552692175, -0.13431550562381744, 0.04758092761039734, 0.004400050267577171, 0.11152182519435883, -0.09896529465913773, 0.07658244669437408, -0.11677825450897217, -0.05863405764102936, 0.0018244348466396332, 0.1012045219540596, -0.27142947912216187, 0.21408241987228394, -0.01260377373546362, -0.03843288496136665, -0.10678957402706146, -0.0025769746862351894, -0.006227991543710232, 0.10789232701063156, 0.10847721993923187, -0.02584187313914299, 0.02608388289809227, -0.0008139219135046005, -0.050714172422885895, 0.037177201360464096, 0.09677128493785858, -0.046560272574424744, 0.0034094313159585, -0.009060622192919254, -0.0012605199590325356, -0.011275713331997395, -0.07949092984199524, -0.025086406618356705, -0.19063347578048706, 0.07284925878047943, 0.0720653235912323, 0.09064801037311554, 0.040668025612831116, -0.04927198961377144, -0.07236650586128235, 0.225349560379982, -0.03647509589791298, -0.10041724145412445, -0.08038459718227386, -0.0284883975982666, 0.04437541216611862, -0.06928593665361404, 0.042854394763708115, -0.03403105214238167, 0.057600777596235275, -0.06608699262142181, -0.14332658052444458, 0.12318196147680283, -0.10255827009677887, -0.041660718619823456, -0.03963806852698326, 0.194327712059021, -0.001014034729450941, 0.005183550529181957, 0.06375695019960403, -0.017277196049690247, -0.10931244492530823, -0.08614889532327652, -0.021600261330604553, 0.04528219252824783, 0.02022574283182621, 0.05683259293437004, -0.056906916201114655, -0.06267181038856506, -0.06268619000911713, -0.03877594321966171, 0.3040119409561157, 0.15447241067886353, -0.02188074216246605, 0.14319118857383728, 0.16920363903045654, -0.08529838174581528, -0.24806492030620575, -0.13262401521205902, -0.059671565890312195, -0.004508363548666239, -0.08310797810554504, -0.17888224124908447, 0.07215721160173416, -0.005662260577082634, -0.014964343048632145, 0.10378111898899078, -0.293643593788147, -0.10397884249687195, 0.16370412707328796, -0.026514273136854172, 0.38681793212890625, -0.08490706980228424, -0.081767737865448, -0.06473196297883987, -0.16839537024497986, 0.1577504426240921, 0.005226342007517815, 0.10632550716400146, -0.012493795715272427, 0.16493095457553864, 0.04338179901242256, 0.02142167091369629, 0.10992388427257538, 0.0510529950261116, -0.05970495194196701, -0.08372924476861954, -0.06349347531795502, -0.02482486516237259, 0.028693553060293198, 0.04520422965288162, -0.04436973109841347, -0.0033934013918042183, -0.14399708807468414, -0.07916028797626495, -0.06317029148340225, 0.044419072568416595, 0.017567407339811325, -0.07359837740659714, 0.0008189964573830366, -0.07994398474693298, 0.0022270381450653076, 0.02018319070339203, 0.14404194056987762, -0.07882563769817352, 0.11366335302591324, 0.0763971209526062, 0.1342843621969223, -0.16993793845176697, -0.015316152945160866, -0.05615348741412163, -0.057570260018110275, 0.0823376253247261, -0.04975086450576782, 0.022654341533780098, 0.12237706035375595, -0.049343910068273544, 0.08374090492725372, 0.10183227062225342, -0.015604394488036633, 0.013752914033830166, 0.10855522751808167, -0.22658367455005646, -0.00024044280871748924, -0.05792419612407684, -0.014340665191411972, 0.10319705307483673, 0.09096817672252655, 0.21301120519638062, 0.0012378974352031946, -0.045622266829013824, -0.006931399926543236, 0.027317756786942482, -0.031812720000743866, 0.08956044912338257, 0.0064537543803453445, -0.0038448970299214125, -0.13701382279396057, 0.05182771384716034, 0.016908686608076096, -0.13509759306907654, 0.02946041151881218, 0.14235343039035797, -0.09227900952100754, -0.12010257691144943, -0.06112139672040939, 0.11130435764789581, -0.18192321062088013, -0.03261134773492813, -0.04455016553401947, -0.13539555668830872, 0.07037829607725143, 0.06719855219125748, 0.06377173960208893, 0.0638071820139885, -0.08599051833152771, 0.000460995826870203, -0.018950268626213074, -0.0022600293159484863, 0.06436912715435028, -0.008528640493750572, -0.05228331685066223, 0.04589683562517166, -0.027857143431901932, 0.10820626467466354, -0.08671467006206512, -0.0914933979511261, -0.14731642603874207, 0.04008224606513977, -0.0835542380809784, -0.06464339792728424, -0.07726794481277466, -0.06062508374452591, -0.0028060092590749264, -0.020913701504468918, -0.048547036945819855, -0.04119202494621277, -0.10599172860383987, 0.04258367419242859, -0.036799151450395584, 0.0242533627897501, -0.06213127076625824, 0.012657810002565384, 0.05118762329220772, -0.04286675155162811, 0.15296420454978943, 0.13311921060085297, -0.0999528095126152, 0.0865686684846878, -0.1682603657245636, -0.04398636892437935, 0.11370586603879929, 0.03027457185089588, 0.06692364066839218, 0.06144743785262108, 0.00934729166328907, 0.06106316298246384, 0.05590246617794037, 0.03528560698032379, 0.017433516681194305, -0.09560048580169678, 0.028305813670158386, -0.04265584796667099, -0.1545007824897766, -0.03425193205475807, -0.020944157615303993, -0.004601715132594109, 0.018733195960521698, 0.07715345174074173, -0.054104406386613846, 0.06715838611125946, -0.04451344534754753, 0.04423696920275688, 0.019160009920597076, -0.14388641715049744, -0.0013057868927717209, -0.0787857249379158, 0.048076074570417404, 0.009943781420588493, 0.20948073267936707, 0.02857227995991707, 0.021133458241820335, 0.010701044462621212, 0.04697025939822197, 0.012060047127306461, 0.012874160893261433, 0.18257758021354675, 0.10245665907859802, -0.02807106263935566, -0.12736539542675018, 0.07029921561479568, 0.06351350247859955, 0.055241890251636505, 0.0994223803281784, -0.055256254971027374, 0.024522610008716583, 0.12849204242229462, 0.0008991300128400326, 0.0211674515157938, -0.1562310755252838, -0.17413091659545898, -0.025765690952539444, 0.05815041437745094, -0.028148863464593887, 0.0932459607720375, 0.1411164551973343, -0.02226310409605503, 0.008420386351644993, -0.008110817521810532, -0.07418473064899445, -0.1858215630054474, -0.1725650131702423, -0.08411741256713867, -0.13248245418071747, 0.0024918527342379093, -0.15276160836219788, 0.03918693959712982, 0.012956192716956139, 0.08509036898612976, -0.07012579590082169, 0.08394156396389008, 0.06268668174743652, -0.11702906340360641, 0.06697677075862885, -0.01844177022576332, 0.10590701550245285, -0.03558329865336418, -0.0026856791228055954, -0.10216444730758667, 0.05418192595243454, -0.0048409937880933285, 0.04099331796169281, -0.045677538961172104, 0.008067529648542404, -0.0895843505859375, -0.05904743820428848, -0.05519740656018257, 0.05829363688826561, -0.00656829122453928, 0.14907732605934143, 0.006240872200578451, -0.04712952673435211, 0.03208540380001068, 0.21914947032928467, -0.06430550664663315, -0.08198840916156769, -0.06634357571601868, 0.19636791944503784, 0.0451631098985672, 0.14286372065544128, -0.0176515094935894, -0.0008699200116097927, -0.10995326936244965, 0.3257746994495392, 0.2731720209121704, -0.0864763930439949, 0.005785810295492411, 0.026608452200889587, 0.04930220544338226, 0.1167016252875328, 0.08703772723674774, 0.11058719456195831, 0.3553784489631653, -0.051051292568445206, -0.019252020865678787, -0.02105122059583664, -0.004171866923570633, -0.08486109226942062, 0.056751929223537445, 0.0487467423081398, -0.05660832300782204, -0.015476187691092491, 0.09990306198596954, -0.21287378668785095, 0.07989609241485596, -0.15139606595039368, -0.1894897222518921, -0.0832260325551033, 0.015674302354454994, 0.0958414375782013, 0.04083721339702606, 0.10753346979618073, -0.015308852307498455, -0.10997200012207031, 0.10574427247047424, 0.02617610991001129, -0.19413700699806213, 0.010398940183222294, 0.07750153541564941, -0.07819200307130814, -0.004374603740870953, -0.025172915309667587, 0.1115877628326416, 0.07077038288116455, 0.06495396047830582, 0.008539075031876564, 0.03771055489778519, 0.002955123782157898, -0.015207790769636631, 0.07194873690605164, 0.028294861316680908, 0.008245851844549179, -0.11345513164997101, 0.07482758909463882, -0.166693776845932, 0.04604841396212578, 0.04979085177183151, -0.028903508558869362, -0.012981443665921688, 0.042016275227069855, -0.08078685402870178, 0.06176526099443436, 0.07577361911535263, -0.007799481973052025, -0.020811740309000015, -0.01000884734094143, 0.011639050208032131, -0.014780595898628235, -0.04551184922456741, -0.07806313037872314, -0.18716010451316833, -0.09903071820735931, 0.0743824690580368, 0.011118550784885883, -0.17065861821174622, 0.0054419320076704025, -0.13119930028915405, 0.05781786888837814, -0.15021705627441406, 0.09188269078731537, 0.07634776085615158, 0.012123112566769123, -0.005486138164997101, 0.002225056756287813, 0.021167593076825142, 0.0780273824930191, -0.13141821324825287, -0.09980349987745285 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.7371 - Matthews Correlation: 0.5481 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.5298 | 1.0 | 535 | 0.5333 | 0.4142 | | 0.3619 | 2.0 | 1070 | 0.5174 | 0.5019 | | 0.2449 | 3.0 | 1605 | 0.6394 | 0.4921 | | 0.1856 | 4.0 | 2140 | 0.7371 | 0.5481 | | 0.133 | 5.0 | 2675 | 0.8600 | 0.5327 | ### Framework versions - Transformers 4.17.0 - Pytorch 1.10.0+cu111 - Datasets 1.18.4 - Tokenizers 0.11.6
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["matthews_correlation"], "model-index": [{"name": "distilbert-base-uncased-finetuned-cola", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "glue", "type": "glue", "args": "cola"}, "metrics": [{"type": "matthews_correlation", "value": 0.5481326292844919, "name": "Matthews Correlation"}]}]}]}
text-classification
BAHIJA/distilbert-base-uncased-finetuned-cola
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-cola ====================================== This model is a fine-tuned version of distilbert-base-uncased on the glue dataset. It achieves the following results on the evaluation set: * Loss: 0.7371 * Matthews Correlation: 0.5481 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5 ### Training results ### Framework versions * Transformers 4.17.0 * Pytorch 1.10.0+cu111 * Datasets 1.18.4 * Tokenizers 0.11.6
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.4\n* Tokenizers 0.11.6" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.4\n* Tokenizers 0.11.6" ]
[ 67, 98, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.17.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.4\n* Tokenizers 0.11.6" ]
[ -0.10176189243793488, 0.09682821482419968, -0.0024596520233899355, 0.12302988767623901, 0.16598010063171387, 0.03423606976866722, 0.13088102638721466, 0.12820081412792206, -0.0833648219704628, 0.021663539111614227, 0.11953437328338623, 0.16124022006988525, 0.023450985550880432, 0.10682792216539383, -0.04690106585621834, -0.2649390399456024, -0.014789818786084652, 0.05157749354839325, -0.05467937886714935, 0.13523271679878235, 0.09001611918210983, -0.12081761658191681, 0.09076643735170364, 0.010078022256493568, -0.19244173169136047, 0.0015631397254765034, 0.0004904005909338593, -0.049425236880779266, 0.14954334497451782, 0.026750387623906136, 0.12263837456703186, 0.004350249655544758, 0.08173038810491562, -0.2013361155986786, 0.010998822748661041, 0.04807887226343155, 0.0019820816814899445, 0.093499556183815, 0.04472099989652634, 0.003356559667736292, 0.12064417451620102, -0.0794675275683403, 0.05245095118880272, 0.0260471161454916, -0.1201077327132225, -0.21434828639030457, -0.07938047498464584, 0.036859773099422455, 0.07537204027175903, 0.10523881763219833, -0.00769936153665185, 0.11958767473697662, -0.0815470889210701, 0.09367476403713226, 0.22499893605709076, -0.2858048677444458, -0.06604119390249252, 0.0446593351662159, 0.013138540089130402, 0.04649971425533295, -0.103193499147892, -0.03450845554471016, 0.0479687862098217, 0.050693217664957047, 0.12551583349704742, -0.02754499390721321, -0.11757545918226242, 0.005937871988862753, -0.1407639980316162, -0.031093906611204147, 0.16633789241313934, 0.04262344166636467, -0.027003895491361618, -0.05469886586070061, -0.05843600630760193, -0.1490705907344818, -0.03513362631201744, -0.01396217755973339, 0.04872048646211624, -0.023977987468242645, -0.040947530418634415, -0.007381310220807791, -0.10852915793657303, -0.06520716100931168, -0.07401566207408905, 0.10994363576173782, 0.03734801709651947, 0.0077004097402095795, -0.028535980731248856, 0.11224483698606491, -0.00789813045412302, -0.12230104207992554, 0.02586428076028824, 0.024746106937527657, 0.012651853263378143, -0.03828500956296921, -0.05250338464975357, -0.06341782957315445, 0.011164400726556778, 0.12632952630519867, -0.056083112955093384, 0.042778655886650085, 0.050611864775419235, 0.050258930772542953, -0.09521278738975525, 0.1895352602005005, -0.034323595464229584, -0.022864026948809624, -0.00014629682118538767, 0.04956270754337311, 0.017138730734586716, -0.011676856316626072, -0.1226736530661583, 0.004377913195639849, 0.08833660930395126, 0.007885524071753025, -0.061885833740234375, 0.07498592138290405, -0.05967307090759277, -0.024604473263025284, 0.004583678673952818, -0.08999699354171753, 0.02237848937511444, 0.0016849944368004799, -0.07211750000715256, -0.022304479032754898, 0.03579258918762207, 0.014694295823574066, -0.0202481672167778, 0.10616497695446014, -0.08772362023591995, 0.028249206021428108, -0.09413883835077286, -0.11074447631835938, 0.017427904531359673, -0.10674558579921722, 0.022673163563013077, -0.091434545814991, -0.17884591221809387, -0.018312914296984673, 0.06050947308540344, -0.024249300360679626, -0.0573410801589489, -0.057426173239946365, -0.06785665452480316, 0.012070219032466412, -0.007625948172062635, 0.11645356565713882, -0.06392399966716766, 0.09426210075616837, 0.02589600719511509, 0.0633111447095871, -0.04242570325732231, 0.05968792736530304, -0.1015826016664505, 0.013873186893761158, -0.1517663300037384, 0.03890359774231911, -0.05208383873105049, 0.07029717415571213, -0.08144880086183548, -0.10629850625991821, 0.004348618909716606, -0.003025123616680503, 0.06253287196159363, 0.09507876634597778, -0.18428830802440643, -0.08226193487644196, 0.1628110557794571, -0.0739215835928917, -0.1214916929602623, 0.1221936047077179, -0.05781576782464981, 0.05870505049824715, 0.05831335484981537, 0.1807072013616562, 0.08820008486509323, -0.07718952745199203, 0.0015782543923705816, 0.02370814047753811, 0.049667228013277054, -0.06470146030187607, 0.06742141395807266, 0.0035519194789230824, 0.0215400829911232, 0.03566690534353256, -0.02571992389857769, 0.06409270316362381, -0.08823181688785553, -0.09829822927713394, -0.03960523009300232, -0.0811794325709343, 0.046684619039297104, 0.08150886744260788, 0.06761076301336288, -0.0943814143538475, -0.07832206785678864, 0.04954811558127403, 0.08154014497995377, -0.057966798543930054, 0.023807577788829803, -0.04936413839459419, 0.07311046123504639, -0.023228125646710396, -0.02161499857902527, -0.1800205409526825, -0.03375808149576187, 0.007829115726053715, 0.002349456073716283, 0.01867217943072319, 0.034778084605932236, 0.06379570066928864, 0.06016054004430771, -0.04999604821205139, -0.019510312005877495, -0.0342816598713398, -0.0011069473112002015, -0.12615589797496796, -0.19688591361045837, -0.030222609639167786, -0.021910198032855988, 0.15960760414600372, -0.20901311933994293, 0.05208921432495117, -0.012877757661044598, 0.06976047158241272, 0.011697076261043549, -0.006303995382040739, -0.038133490830659866, 0.07676824182271957, -0.04173412173986435, -0.050108104944229126, 0.08339998126029968, 0.014047103933990002, -0.09057528525590897, -0.04945938661694527, -0.09853032976388931, 0.15839216113090515, 0.12883667647838593, -0.1116182953119278, -0.07826980203390121, -0.02305111661553383, -0.06671160459518433, -0.03522639721632004, -0.04416444152593613, 0.025876345112919807, 0.18832464516162872, -0.004962592851370573, 0.14948441088199615, -0.06832242012023926, -0.04274310916662216, 0.019425585865974426, -0.036795347929000854, 0.017852047458291054, 0.13590803742408752, 0.13378944993019104, -0.06104964762926102, 0.15585274994373322, 0.14639239013195038, -0.08337175846099854, 0.15007604658603668, -0.04265443980693817, -0.06496241688728333, -0.016440503299236298, -0.029011590406298637, -0.010684406384825706, 0.10029394179582596, -0.1562347561120987, -0.001848764717578888, 0.030182138085365295, 0.015586490742862225, 0.025542765855789185, -0.22884155809879303, -0.0414334274828434, 0.0374080128967762, -0.0445118173956871, -0.008143243379890919, -0.006773552857339382, 0.004530115518718958, 0.10129335522651672, -0.0006000894936732948, -0.08817089349031448, 0.03657932206988335, 0.0025305224116891623, -0.08364532142877579, 0.21628373861312866, -0.07994058728218079, -0.170803964138031, -0.13348102569580078, -0.07103874534368515, -0.04671558737754822, -0.0010986656416207552, 0.06840936094522476, -0.09734582901000977, -0.025647521018981934, -0.07164523005485535, 0.02599717676639557, 0.008444010280072689, 0.02195236273109913, 0.0026903694961220026, 0.007644058670848608, 0.06510692834854126, -0.11232278496026993, -0.014881678856909275, -0.05841048061847687, -0.04633007198572159, 0.04644492268562317, 0.027525408193469048, 0.1094992384314537, 0.15353281795978546, -0.012758109718561172, 0.011892959475517273, -0.030734224244952202, 0.23498183488845825, -0.06016358733177185, -0.019647788256406784, 0.14621403813362122, -0.008030868135392666, 0.05282779783010483, 0.11456439644098282, 0.07550833374261856, -0.07688812911510468, 0.003238122910261154, 0.03863020986318588, -0.03525501862168312, -0.2327682226896286, -0.05309433862566948, -0.054635096341371536, 0.013064105994999409, 0.08935478329658508, 0.02416999638080597, 0.029937075451016426, 0.07054702192544937, 0.04136786237359047, 0.07332197576761246, -0.03811226412653923, 0.049971457570791245, 0.13011164963245392, 0.029262414202094078, 0.12461420148611069, -0.046886224299669266, -0.06492874771356583, 0.03935838118195534, -0.010637765750288963, 0.22409991919994354, 0.009192897006869316, 0.13297748565673828, 0.06587589532136917, 0.16391873359680176, -0.00994152668863535, 0.07602468132972717, -0.011395679786801338, -0.038871705532073975, -0.015606730245053768, -0.03918904438614845, -0.03953630477190018, 0.02350275218486786, -0.06447871774435043, 0.0643714889883995, -0.12362480163574219, 0.013744720257818699, 0.058600302785634995, 0.24754568934440613, 0.035082850605249405, -0.3189503848552704, -0.09865427762269974, 0.00042755878530442715, -0.03040378913283348, -0.018965424969792366, 0.02581312321126461, 0.09616107493638992, -0.09647485613822937, 0.027882108464837074, -0.07374172657728195, 0.09709927439689636, -0.055670130997896194, 0.05182754248380661, 0.08212333917617798, 0.0890168696641922, 0.012233886867761612, 0.09444060176610947, -0.2862294018268585, 0.2772308588027954, 0.0003452537639532238, 0.05551348626613617, -0.07580750435590744, 0.00853610597550869, 0.04082805663347244, 0.06510092318058014, 0.07884475588798523, -0.011240086518228054, -0.018044855445623398, -0.1894557923078537, -0.06712591648101807, 0.026666143909096718, 0.0684526115655899, -0.040632978081703186, 0.08300479501485825, -0.03066869266331196, 0.008531827479600906, 0.07304056733846664, 0.0036414831411093473, -0.05114342272281647, -0.1069520115852356, -0.00489991856738925, 0.02188265509903431, -0.061132002621889114, -0.06084221601486206, -0.12109033763408661, -0.12898151576519012, 0.15592463314533234, -0.033967524766922, -0.03862280771136284, -0.10550474375486374, 0.08371008932590485, 0.059271495789289474, -0.09044332057237625, 0.04167364537715912, 0.002501430455595255, 0.07362639904022217, 0.02123427204787731, -0.07079128921031952, 0.10443390905857086, -0.07293680310249329, -0.1556394100189209, -0.06561664491891861, 0.10699650645256042, 0.03266165405511856, 0.06692623347043991, -0.014606797136366367, 0.004896017722785473, -0.04696812480688095, -0.08834145963191986, 0.022342724725604057, 0.000684437807649374, 0.07681337743997574, 0.01882920414209366, -0.07533686608076096, 0.014076688326895237, -0.06499658524990082, -0.03383789211511612, 0.20455153286457062, 0.22098788619041443, -0.09996180236339569, 0.022009236738085747, 0.026295019313693047, -0.07341528683900833, -0.1981097161769867, 0.0341200977563858, 0.05544741824269295, 0.008141477592289448, 0.04227812588214874, -0.18522965908050537, 0.1312389224767685, 0.10763580352067947, -0.01162867434322834, 0.10807356983423233, -0.32348737120628357, -0.11982133239507675, 0.1347520798444748, 0.1356658935546875, 0.10056072473526001, -0.13414393365383148, -0.022965392097830772, -0.017363306134939194, -0.14047491550445557, 0.11419897526502609, -0.09213969111442566, 0.12174172699451447, -0.03840717673301697, 0.07624481618404388, 0.002732468768954277, -0.05879552289843559, 0.11987645924091339, 0.025696398690342903, 0.0950632318854332, -0.05975884571671486, -0.034254372119903564, 0.031001843512058258, -0.042627569288015366, 0.033712439239025116, -0.09849841147661209, 0.02948264218866825, -0.10483381152153015, -0.024378858506679535, -0.06985113024711609, 0.04517354071140289, -0.04560216888785362, -0.06741434335708618, -0.03660593554377556, 0.026076875627040863, 0.0447082556784153, -0.007506086491048336, 0.12256038933992386, 0.023544639348983765, 0.14991986751556396, 0.10042021423578262, 0.07233428955078125, -0.06990049034357071, -0.0826539620757103, -0.028090139850974083, -0.010032063350081444, 0.050856392830610275, -0.1381862461566925, 0.01939801685512066, 0.15314216911792755, 0.02105153724551201, 0.1535300761461258, 0.08343400806188583, -0.022126782685518265, -0.0006706930580548942, 0.05952814221382141, -0.16435027122497559, -0.09419874101877213, -0.016728220507502556, -0.06860101222991943, -0.119585320353508, 0.04558093473315239, 0.09271853417158127, -0.06780552119016647, -0.006338559556752443, -0.004187530837953091, 0.013843416236341, -0.05063757672905922, 0.18390674889087677, 0.06337220966815948, 0.04899980127811432, -0.09597993642091751, 0.07034490257501602, 0.04625708982348442, -0.07393568009138107, 0.0018211835995316505, 0.07275162637233734, -0.08483824878931046, -0.055753376334905624, 0.06379304826259613, 0.18950462341308594, -0.044147126376628876, -0.04757954925298691, -0.14624649286270142, -0.12364145368337631, 0.07680732756853104, 0.14099369943141937, 0.11872129142284393, 0.011112731881439686, -0.0656782016158104, 0.002607102505862713, -0.10715238749980927, 0.10063575953245163, 0.044938504695892334, 0.06259337067604065, -0.142737478017807, 0.1424248069524765, 0.0207118708640337, 0.04932249337434769, -0.01794695481657982, 0.02275070920586586, -0.10082220286130905, 0.0073870643973350525, -0.09323610365390778, -0.020332418382167816, -0.029273292049765587, 0.0115242013707757, -0.006042692810297012, -0.046746160835027695, -0.053276944905519485, 0.010670280084013939, -0.10750047862529755, -0.023516593500971794, 0.03070518560707569, 0.07236143201589584, -0.1089014932513237, -0.036667175590991974, 0.03138185665011406, -0.0602446123957634, 0.07398806512355804, 0.043224018067121506, 0.01618783175945282, 0.05050251632928848, -0.1388249546289444, 0.0205977950245142, 0.07254750281572342, 0.029431859031319618, 0.06116562709212303, -0.09837362170219421, -0.007884427905082703, -0.009744220413267612, 0.039233673363924026, 0.021287856623530388, 0.0740664154291153, -0.14142771065235138, 0.004113059025257826, -0.02347916178405285, -0.08444929867982864, -0.06769412755966187, 0.028111180290579796, 0.08742963522672653, 0.017817210406064987, 0.1985483020544052, -0.07610398530960083, 0.05088236182928085, -0.2202623039484024, 0.0067904917523264885, -0.005604913458228111, -0.11054443567991257, -0.09816008806228638, -0.07262846827507019, 0.05472603440284729, -0.06091339513659477, 0.1482498198747635, 0.046197887510061264, 0.01825208030641079, 0.02316269651055336, -0.0110246567055583, 0.011770087294280529, 0.00974032562226057, 0.18850935995578766, 0.03009743243455887, -0.03421899303793907, 0.05876551568508148, 0.04535519331693649, 0.10122247785329819, 0.11457179486751556, 0.1984933763742447, 0.14438314735889435, -0.009702610783278942, 0.09137076139450073, 0.04536355286836624, -0.055524811148643494, -0.155462846159935, 0.05141647160053253, -0.0356324128806591, 0.1088876724243164, -0.021074756979942322, 0.2187509983778, 0.06435628980398178, -0.16816246509552002, 0.052556298673152924, -0.05240139365196228, -0.0878315269947052, -0.1135190799832344, -0.047323133796453476, -0.07605129480361938, -0.1318512111902237, -0.0037816669791936874, -0.11554285883903503, -0.003295320551842451, 0.12721122801303864, 0.0041609713807702065, -0.027361437678337097, 0.1592894345521927, 0.01586962305009365, 0.023365488275885582, 0.058633312582969666, 0.00845402106642723, -0.03855054825544357, -0.13824526965618134, -0.058141861110925674, -0.011803640983998775, -0.009254679083824158, 0.030130017548799515, -0.06108370050787926, -0.04366555064916611, 0.03079751878976822, -0.02023622952401638, -0.09625943750143051, 0.005226355977356434, 0.011334143579006195, 0.05347541719675064, 0.045377980917692184, 0.009317897260189056, 0.01945488341152668, -0.004537243861705065, 0.1998896449804306, -0.07096187025308609, -0.06452770531177521, -0.10131637006998062, 0.23608043789863586, 0.03418838605284691, -0.018943285569548607, 0.03344103321433067, -0.06629064679145813, 0.004891349468380213, 0.24789069592952728, 0.2161741703748703, -0.08045034110546112, -0.005819495767354965, 0.01792411506175995, -0.007871708832681179, -0.02049819380044937, 0.09874389320611954, 0.1428508758544922, 0.04796459525823593, -0.09266992658376694, -0.044615488499403, -0.05907188355922699, -0.01757912151515484, -0.034729570150375366, 0.06913961470127106, 0.05202385410666466, 0.010288444347679615, -0.03627867251634598, 0.056454479694366455, -0.06655225902795792, -0.08974503725767136, 0.05961492657661438, -0.2199363261461258, -0.166675865650177, -0.016223229467868805, 0.10351404547691345, 0.002952094655483961, 0.062420692294836044, -0.02990744635462761, -0.0029332435224205256, 0.08977600932121277, -0.019349532201886177, -0.09828799217939377, -0.07286416739225388, 0.08377633988857269, -0.11333484202623367, 0.2166394591331482, -0.048994600772857666, 0.055853698402643204, 0.12577374279499054, 0.06741653382778168, -0.06447659432888031, 0.06542164087295532, 0.04164041206240654, -0.042867425829172134, 0.022093839943408966, 0.06810236722230911, -0.03162214905023575, 0.06304416805505753, 0.047801896929740906, -0.13935768604278564, 0.02326846681535244, -0.04629216343164444, -0.0686112493276596, -0.044165175408124924, -0.021770309656858444, -0.05967537686228752, 0.12940135598182678, 0.22077418863773346, -0.024954870343208313, -0.010420500300824642, -0.07177383452653885, 0.006767777260392904, 0.05553433299064636, 0.020797887817025185, -0.05836331471800804, -0.2100914567708969, 0.01786745712161064, 0.045134179294109344, -0.017657319083809853, -0.25293514132499695, -0.10050103813409805, 0.004428518936038017, -0.07400961220264435, -0.09425722062587738, 0.07090505957603455, 0.08793305605649948, 0.055241793394088745, -0.05599416419863701, -0.048706162720918655, -0.0744684487581253, 0.14993782341480255, -0.14478886127471924, -0.09266992658376694 ]
null
null
transformers
[Paper](https://hlt.bme.hu/en/publ/foszt2oszt) We publish an abstractive summarizer for Hungarian, an encoder-decoder model initialized with [huBERT](huggingface.co/SZTAKI-HLT/hubert-base-cc), and fine-tuned on the [ELTE.DH](https://elte-dh.hu/) corpus of former Hungarian news portals. The model produces fluent output in the correct topic, but it hallucinates frequently. Our quantitative evaluation on automatic and human transcripts of news (with automatic and human-made punctuation, [Tündik et al. (2019)](https://www.isca-speech.org/archive/interspeech_2019/tundik19_interspeech.html), [Tündik and Szaszák (2019)](https://www.isca-speech.org/archive/interspeech_2019/szaszak19_interspeech.html)) shows that the model is robust with respect to errors in either automatic speech recognition or automatic punctuation restoration. In fine-tuning and inference, we followed [a jupyter notebook by Patrick von Platen](https://github.com/patrickvonplaten/notebooks/blob/master/BERT2BERT_for_CNN_Dailymail.ipynb). Most hyper-parameters are the same as those by von Platen, but we found it advantageous to change the minimum length of the summary to 8 word- pieces (instead of 56), and the number of beams in beam search to 5 (instead of 4). Our model was fine-tuned on a server of the [SZTAKI-HLT](hlt.bme.hu/) group, which kindly provided access to it.
{"language": "hu", "metrics": "rouge"}
text2text-generation
BME-TMIT/foszt2oszt
[ "transformers", "pytorch", "encoder-decoder", "text2text-generation", "hu", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[ "hu" ]
TAGS #transformers #pytorch #encoder-decoder #text2text-generation #hu #autotrain_compatible #endpoints_compatible #region-us
Paper We publish an abstractive summarizer for Hungarian, an encoder-decoder model initialized with huBERT, and fine-tuned on the ELTE.DH corpus of former Hungarian news portals. The model produces fluent output in the correct topic, but it hallucinates frequently. Our quantitative evaluation on automatic and human transcripts of news (with automatic and human-made punctuation, Tündik et al. (2019), Tündik and Szaszák (2019)) shows that the model is robust with respect to errors in either automatic speech recognition or automatic punctuation restoration. In fine-tuning and inference, we followed a jupyter notebook by Patrick von Platen. Most hyper-parameters are the same as those by von Platen, but we found it advantageous to change the minimum length of the summary to 8 word- pieces (instead of 56), and the number of beams in beam search to 5 (instead of 4). Our model was fine-tuned on a server of the SZTAKI-HLT group, which kindly provided access to it.
[]
[ "TAGS\n#transformers #pytorch #encoder-decoder #text2text-generation #hu #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 44 ]
[ "passage: TAGS\n#transformers #pytorch #encoder-decoder #text2text-generation #hu #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.04170882701873779, 0.0724821463227272, -0.0062649487517774105, 0.0001537116913823411, 0.18010886013507843, 0.01650948077440262, 0.08875813335180283, 0.1252685785293579, -0.02026052214205265, -0.009348861873149872, 0.08236560225486755, 0.1587824672460556, 0.016807518899440765, 0.13752736151218414, -0.08737515658140182, -0.2281450480222702, 0.07087717205286026, 0.06971722841262817, 0.09407956153154373, 0.09927612543106079, 0.0654165968298912, -0.07298903912305832, 0.10573229938745499, -0.03908232972025871, -0.13107633590698242, 0.05232284590601921, 0.00450938381254673, -0.09052275866270065, 0.07219308614730835, 0.007245566695928574, 0.1303717941045761, 0.032846152782440186, -0.012711231596767902, -0.16944198310375214, 0.019655268639326096, 0.01660461165010929, -0.05484496429562569, 0.02830522507429123, 0.07524038106203079, -0.05796395242214203, 0.0763966515660286, 0.06290512531995773, -0.044388968497514725, 0.07065165787935257, -0.10767233371734619, -0.04564960300922394, -0.02592252753674984, 0.04835965856909752, 0.06746140122413635, 0.09051837772130966, -0.011383178643882275, 0.1106332540512085, -0.08811285346746445, 0.1262548416852951, 0.11257094889879227, -0.26357337832450867, 0.025779306888580322, 0.0205596424639225, 0.0960727110505104, 0.10075780004262924, -0.0320536345243454, 0.027426397427916527, 0.04773417115211487, 0.050187405198812485, -0.034893136471509933, -0.07233765721321106, -0.17397324740886688, 0.009074973873794079, -0.054677050560712814, -0.04689266160130501, 0.19385482370853424, -0.07152225822210312, 0.06866755336523056, -0.022670051082968712, -0.10684198886156082, -0.1257370561361313, -0.03448863327503204, 0.04768501594662666, -0.06317617744207382, 0.08054456114768982, 0.03640732169151306, -0.023059526458382607, -0.10888046026229858, 0.011559347622096539, -0.17648877203464508, 0.13930897414684296, 0.02904362790286541, 0.05246812105178833, -0.19408774375915527, 0.07707106322050095, 0.1627533882856369, -0.14838860929012299, 0.04671664908528328, -0.06908927112817764, 0.07406117767095566, 0.008561030961573124, -0.06786099076271057, -0.04548776149749756, 0.08685249090194702, 0.09032431989908218, -0.048634786158800125, 0.06625194102525711, -0.04900987073779106, 0.09548463672399521, 0.002444304758682847, 0.09613219648599625, 0.013812447898089886, -0.04512135311961174, 0.046107541769742966, -0.09733432531356812, 0.03660164400935173, -0.06823185831308365, -0.16989855468273163, -0.02854818105697632, 0.0522640161216259, 0.07875855267047882, 0.009282398968935013, 0.051810603588819504, -0.05878722667694092, 0.005296456161886454, 0.05275756120681763, -0.078524149954319, 0.029915399849414825, 0.024139096960425377, 0.035845592617988586, 0.1367509514093399, 0.004326289985328913, 0.01622038520872593, -0.15233175456523895, 0.05935634300112724, -0.04907606542110443, 0.033408697694540024, -0.04321649670600891, -0.06607368588447571, 0.03530372679233551, -0.1472288817167282, 0.013474109582602978, -0.14592492580413818, -0.10864409059286118, 0.00461495341733098, 0.0014623608440160751, 0.007983335293829441, -0.009036912582814693, -0.04130956158041954, -0.029018670320510864, 0.04388894513249397, -0.08052905648946762, -0.12576152384281158, -0.04640273377299309, 0.09792438894510269, -0.02132091112434864, 0.09305274486541748, -0.16870588064193726, 0.07351715862751007, -0.1197044625878334, -0.008561724796891212, -0.12259834259748459, 0.06833956390619278, -0.02157476544380188, 0.17972294986248016, 0.020011121407151222, -0.04730228707194328, -0.05465030297636986, 0.0453527458012104, -0.011120208539068699, 0.1477009505033493, -0.08551900833845139, -0.10337647050619125, 0.24098999798297882, -0.11470102518796921, -0.11580639332532883, 0.11220850795507431, 0.028802597895264626, -0.03516680374741554, 0.06518133729696274, 0.19825927913188934, 0.07445461302995682, -0.01321902871131897, 0.037931010127067566, 0.15164220333099365, -0.08422303944826126, -0.11478634923696518, -0.015223727561533451, -0.015814006328582764, -0.04639287665486336, 0.04693785309791565, 0.1090884804725647, 0.10966137796640396, -0.04871100187301636, -0.053644806146621704, -0.028359180316329002, -0.006658956408500671, 0.06199817731976509, 0.04005562141537666, 0.09748191386461258, -0.07694951444864273, -0.0065210238099098206, -0.04137285798788071, 0.009121823124587536, -0.03605138510465622, 0.07208655029535294, -0.06270012259483337, 0.10683979839086533, 0.007444518152624369, 0.03300517052412033, -0.1569453328847885, -0.02691449224948883, -0.018611811101436615, 0.08682030439376831, -0.045543819665908813, 0.05277221277356148, 0.030494071543216705, -0.0024532561656087637, 0.0035876529291272163, -0.043593402951955795, 0.11482400447130203, 0.006348741706460714, -0.04608170688152313, -0.052261799573898315, 0.022822314873337746, -0.0418822318315506, 0.00748703395947814, -0.03193964809179306, 0.021943749859929085, 0.039290472865104675, 0.177407905459404, -0.0007871125708334148, 0.03491726145148277, -0.0349486880004406, 0.0941891148686409, -0.09606077522039413, 0.013913258910179138, 0.0969453677535057, 0.002637293888255954, -0.06481262296438217, 0.20159609615802765, -0.15334293246269226, 0.21117059886455536, 0.21156392991542816, -0.2928462326526642, 0.041728243231773376, 0.03314938396215439, -0.004667756147682667, 0.011648650281131268, 0.055242445319890976, -0.016141941770911217, 0.12630786001682281, -0.009855994023382664, 0.21780841052532196, -0.03463524952530861, -0.010922993533313274, -0.005694216582924128, -0.06730151921510696, -0.009699035435914993, 0.05222811922430992, 0.08532895892858505, -0.11415835469961166, 0.16688275337219238, 0.15255694091320038, 0.0036902371793985367, 0.19020669162273407, -0.02939566969871521, -0.017353324219584465, 0.09590547531843185, -0.001921147690154612, -0.050793904811143875, -0.06797826290130615, -0.16626884043216705, -0.03316351771354675, 0.05215095356106758, 0.037444744259119034, 0.12015500664710999, -0.12202250957489014, -0.0038448721170425415, -0.031777091324329376, -0.0048357113264501095, -0.037041809409856796, 0.06648501753807068, 0.07036004215478897, 0.0821596309542656, -0.04212133213877678, -0.027860596776008606, 0.0920688733458519, 0.010080144740641117, -0.09831461310386658, 0.14371640980243683, -0.17004865407943726, -0.3490133583545685, -0.23238687217235565, -0.15791203081607819, -0.0193022470921278, 0.050469767302274704, 0.13908414542675018, -0.11561081558465958, -0.04368795454502106, 0.00037170140421949327, 0.01273664552718401, -0.024801308289170265, -0.012522065080702305, 0.04794749990105629, 0.058275848627090454, -0.014950886368751526, -0.06955952197313309, -0.04447600245475769, 0.022567644715309143, 0.003272389993071556, 0.14233197271823883, -0.10279621928930283, 0.10069450736045837, 0.14605771005153656, 0.008528286591172218, 0.07299058139324188, -0.006112322211265564, 0.12922866642475128, -0.06496378779411316, -0.028605422005057335, 0.22382648289203644, -0.08937712758779526, 0.08929554373025894, 0.14434196054935455, -0.009054348804056644, -0.05075718089938164, 0.005453686695545912, -0.07154473662376404, -0.07205288857221603, -0.16590063273906708, -0.146641343832016, -0.13521762192249298, 0.055568769574165344, 0.024432072415947914, 0.05843213200569153, 0.06062190607190132, 0.05503349378705025, -0.027370823547244072, 0.016115626320242882, -0.0375397652387619, 0.07392498850822449, 0.20004405081272125, -0.018541114404797554, 0.11642924696207047, -0.08400124311447144, -0.11789808422327042, 0.08853968232870102, 0.061403289437294006, 0.08384174108505249, 0.04827491566538811, 0.05891328677535057, 0.003538004355505109, 0.07939814776182175, 0.14557941257953644, 0.10994856804609299, 0.03231116756796837, -0.004396187607198954, 0.015578165650367737, -0.0027553264517337084, -0.052344609051942825, 0.008977158926427364, 0.007916917093098164, -0.13185414671897888, -0.08388566225767136, -0.024854741990566254, 0.09561249613761902, 0.0875796303153038, 0.08419982343912125, -0.24361900985240936, 0.010210203938186169, 0.08718591928482056, -0.015311400406062603, -0.12321159988641739, 0.09202385693788528, 0.036276139318943024, -0.09837477654218674, 0.0872301459312439, 0.005489995237439871, 0.13106292486190796, -0.0692400112748146, 0.08337270468473434, -0.03542136400938034, -0.05322534963488579, 0.05741574242711067, 0.09711112827062607, -0.2325233370065689, 0.14399145543575287, -0.022502819076180458, -0.028654620051383972, -0.05966585874557495, 0.008908652700483799, 0.02581772394478321, 0.1691412478685379, 0.055989820510149, 0.01798400841653347, -0.05563950538635254, -0.18006198108196259, -0.03176293149590492, 0.030056586489081383, 0.14046017825603485, 0.02542884461581707, 0.010761885903775692, -0.05102995038032532, -0.01767553947865963, -0.058916181325912476, -0.08687140792608261, 0.00029584765434265137, -0.1631074994802475, 0.07311845570802689, 0.13319313526153564, 0.11723321676254272, -0.0018329614540562034, -0.024061089381575584, -0.10965969413518906, 0.17923729121685028, -0.06312920898199081, -0.10044234991073608, -0.09900600463151932, -0.10412905365228653, 0.037887509912252426, -0.07236179709434509, 0.040131792426109314, -0.0523822121322155, 0.01911459118127823, -0.06302352994680405, -0.19479292631149292, 0.09973206371068954, -0.09733158349990845, -0.023514406755566597, -0.03806263580918312, 0.15933789312839508, -0.08174245804548264, -0.008117246441543102, 0.021512599661946297, -0.008286233991384506, -0.10075490921735764, -0.09244004637002945, -0.05256207659840584, 0.0038603972643613815, 0.08346252888441086, 0.07446005195379257, -0.08716640621423721, -0.03210000693798065, -0.010579998604953289, -0.015054657123982906, 0.2823999524116516, 0.09748119860887527, -0.05669720843434334, 0.16778771579265594, 0.15239068865776062, -0.06628528237342834, -0.3650761544704437, -0.15272875130176544, -0.09808672219514847, -0.016800733283162117, -0.05966858193278313, -0.14412514865398407, 0.12029892206192017, -0.05524572730064392, -0.009181096218526363, 0.09687652438879013, -0.20711958408355713, -0.05756460502743721, 0.15341489017009735, -0.019476257264614105, 0.30251333117485046, -0.08742973208427429, -0.08457767218351364, -0.0773005560040474, -0.2541532814502716, 0.017476936802268028, -0.043739959597587585, 0.08216521143913269, -0.01908668875694275, 0.08926659822463989, 0.03324282541871071, -0.05215463414788246, 0.08490968495607376, 0.00764976441860199, -0.005919781979173422, -0.1155606135725975, 0.009587748907506466, 0.09185421466827393, -0.0065976399928331375, 0.033462781459093094, -0.12105820327997208, 0.046034444123506546, -0.15252548456192017, -0.013122275471687317, -0.07419716566801071, 0.04003547132015228, 0.03609875962138176, -0.009837053716182709, 0.01030933391302824, -0.0686536431312561, 0.038231078535318375, 0.03165510669350624, 0.17921072244644165, -0.03355075791478157, 0.12617379426956177, 0.144414022564888, 0.11609697341918945, -0.21792833507061005, 0.07684401422739029, -0.06317243725061417, -0.07363693416118622, 0.07679495960474014, 0.006548753008246422, 0.08394209295511246, 0.12853918969631195, -0.06945164501667023, 0.06524290889501572, 0.1042722538113594, 0.05932354927062988, 0.020824363455176353, 0.16428759694099426, -0.2239040583372116, 0.03790964558720589, -0.09111151844263077, -0.05870753154158592, 0.06615174561738968, 0.08892679214477539, 0.15864810347557068, 0.043314337730407715, -0.027107587084174156, -0.024753043428063393, 0.021252259612083435, -0.05987091362476349, 0.04414980486035347, 0.004285712726414204, 0.03140128031373024, -0.16905109584331512, 0.05567723512649536, 0.004115339368581772, -0.13485926389694214, 0.02168053388595581, 0.19973523914813995, -0.17272691428661346, -0.1167430505156517, -0.04403407499194145, 0.18269795179367065, -0.10728264600038528, -0.07875064015388489, -0.10359317809343338, -0.1411897838115692, 0.05357323959469795, 0.13402877748012543, 0.07695283740758896, 0.07077858597040176, -0.03980930149555206, -0.03647993132472038, -0.05951617285609245, -0.004921888932585716, 0.02503424882888794, 0.042335450649261475, -0.13368451595306396, 0.08997491002082825, -0.027649926021695137, 0.14941538870334625, -0.08767911791801453, -0.040271732956171036, -0.12606534361839294, 0.032393574714660645, -0.19017188251018524, -0.049985047429800034, -0.09779723733663559, -0.03871426731348038, -0.009500887244939804, -0.035201940685510635, -0.05824219062924385, -0.013941026292741299, -0.09282458573579788, -0.014137252233922482, -0.0400206483900547, 0.02273331582546234, -0.06995845586061478, 0.020733891054987907, 0.07443546503782272, -0.048329535871744156, 0.11651811003684998, 0.12123569846153259, -0.10521110147237778, 0.09117135405540466, -0.06895095854997635, -0.15843921899795532, 0.13669778406620026, 0.030502378940582275, 0.04505625367164612, 0.11988463997840881, 0.01421030331403017, 0.09732405096292496, 0.007676284294575453, 0.030824994668364525, 0.10680844634771347, -0.08802742511034012, 0.019679391756653786, -0.08538966625928879, -0.10991286486387253, -0.06144517660140991, -0.011839404702186584, 0.09974261373281479, 0.01404134463518858, 0.13420897722244263, -0.05677535757422447, 0.1001504436135292, -0.04020053520798683, 0.005838148295879364, 0.0038386608939617872, -0.15142185986042023, -0.06201530620455742, -0.10060539096593857, 0.02695482224225998, -0.013906835578382015, 0.22324776649475098, 0.0060707200318574905, 0.019138505682349205, 0.021181171759963036, 0.06678759306669235, -0.06884104758501053, 0.012947817333042622, 0.24165946245193481, 0.08443192392587662, -0.07180911302566528, -0.11109312623739243, 0.0284594539552927, 0.028694922104477882, -0.008693643845617771, 0.1075887456536293, 0.06751323491334915, -0.01060936227440834, 0.09742984920740128, -0.01862313784658909, 0.025399954989552498, -0.19122104346752167, -0.12946216762065887, -0.06329543143510818, 0.07274819910526276, 0.013804621994495392, 0.0878787636756897, 0.14958854019641876, -0.006383171305060387, 0.023374952375888824, -0.006587490439414978, -0.04226289317011833, -0.1597871035337448, -0.06433399021625519, -0.0801156535744667, -0.16741473972797394, 0.02046819217503071, -0.07784552127122879, 0.03647690638899803, 0.03191960230469704, 0.021191665902733803, -0.07367052882909775, 0.13514705002307892, 0.05190534517168999, -0.09040182828903198, 0.04367772862315178, -0.03881467878818512, 0.06525278091430664, 0.04084549844264984, -0.048225272446870804, -0.057569295167922974, -0.013855226337909698, 0.007458777632564306, 0.0455627478659153, -0.05589429661631584, 0.036259252578020096, -0.1488247662782669, -0.10954871028661728, -0.01419107150286436, 0.04393643140792847, -0.03013506531715393, 0.11976740509271622, 0.02914937399327755, -0.011787456460297108, 0.01769215427339077, 0.14361393451690674, -0.09240416437387466, -0.10472869873046875, -0.04690228030085564, 0.20482446253299713, 0.0915168896317482, 0.08779305964708328, -0.010916895233094692, -0.018459029495716095, -0.12271928042173386, 0.33078745007514954, 0.23929570615291595, -0.08673355728387833, 0.019808156415820122, 0.015131540596485138, 0.03838846832513809, 0.08879998326301575, 0.12939469516277313, 0.12502698600292206, 0.2384139448404312, -0.04041733965277672, -0.06258556991815567, -0.056888312101364136, -0.04203824698925018, -0.16603587567806244, 0.06554818898439407, -0.0016079371562227607, -0.08026617765426636, -0.049169912934303284, 0.06278092414140701, -0.21419937908649445, 0.16195492446422577, -0.02417161501944065, -0.19614748656749725, -0.007232749368995428, -0.004786001518368721, 0.22222928702831268, 0.009543281979858875, 0.02734226919710636, 0.006465838756412268, -0.08498326689004898, 0.1384485363960266, 0.0002664573257789016, -0.17458295822143555, 0.04662678763270378, -0.002659640507772565, -0.25603029131889343, -0.02311168611049652, -0.034271400421857834, 0.017629405483603477, 0.07456909865140915, 0.10617472976446152, -0.019312411546707153, 0.09086337685585022, -0.028180554509162903, -0.03389307111501694, 0.0059915692545473576, 0.06824269145727158, -0.008345436304807663, -0.0918019488453865, 0.04250473901629448, -0.14558042585849762, 0.021551022306084633, -0.05925006791949272, -0.018510347232222557, 0.02993139624595642, -0.0397813655436039, -0.028742343187332153, 0.06378711014986038, 0.06146784499287605, 0.0027328284922987223, 0.008425462990999222, -0.03482751175761223, -0.0339014045894146, -0.003583083627745509, -0.10741734504699707, -0.08178610354661942, -0.12831509113311768, -0.11871061474084854, 0.12162695080041885, 0.029857760295271873, -0.14763307571411133, 0.028600240126252174, -0.12119197100400925, 0.04860756918787956, -0.19177275896072388, 0.08839421719312668, 0.09184100478887558, 0.009049090556800365, -0.000492636754643172, -0.05331961438059807, 0.059037331491708755, 0.09477448463439941, -0.11441460251808167, -0.10628295689821243 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/RoBERTalex # Spanish Legal-domain RoBERTa There are few models trained for the Spanish language. Some of the models have been trained with a low resource, unclean corpora. The ones derived from the Spanish National Plan for Language Technologies are proficient solving several tasks and have been trained using large scale clean corpora. However, the Spanish Legal domain language could be think of an independent language on its own. We therefore created a Spanish Legal model from scratch trained exclusively on legal corpora. ## Citing ``` @misc{gutierrezfandino2021legal, title={Spanish Legalese Language Model and Corpora}, author={Asier Gutiérrez-Fandiño and Jordi Armengol-Estapé and Aitor Gonzalez-Agirre and Marta Villegas}, year={2021}, eprint={2110.12201}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` For more information visit our [GitHub repository](https://github.com/PlanTL-GOB-ES/lm-legal-es) ## Funding This work was funded by the Spanish State Secretariat for Digitalization and Artificial Intelligence (SEDIA) within the framework of the Plan-TL.
{"language": ["es"], "license": "apache-2.0", "tags": ["legal", "spanish"], "datasets": ["legal_ES", "temu_legal"], "metrics": ["ppl"], "widget": [{"text": "La ley fue <mask> finalmente."}, {"text": "El Tribunal <mask> desestim\u00f3 el recurso de amparo."}, {"text": "Hay base legal dentro del marco <mask> actual."}]}
fill-mask
BSC-LT/RoBERTalex
[ "transformers", "pytorch", "roberta", "fill-mask", "legal", "spanish", "es", "dataset:legal_ES", "dataset:temu_legal", "arxiv:2110.12201", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "2110.12201" ]
[ "es" ]
TAGS #transformers #pytorch #roberta #fill-mask #legal #spanish #es #dataset-legal_ES #dataset-temu_legal #arxiv-2110.12201 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL # Spanish Legal-domain RoBERTa There are few models trained for the Spanish language. Some of the models have been trained with a low resource, unclean corpora. The ones derived from the Spanish National Plan for Language Technologies are proficient solving several tasks and have been trained using large scale clean corpora. However, the Spanish Legal domain language could be think of an independent language on its own. We therefore created a Spanish Legal model from scratch trained exclusively on legal corpora. ## Citing For more information visit our GitHub repository ## Funding This work was funded by the Spanish State Secretariat for Digitalization and Artificial Intelligence (SEDIA) within the framework of the Plan-TL.
[ "# Spanish Legal-domain RoBERTa\n\nThere are few models trained for the Spanish language. Some of the models have been trained with a low resource, unclean corpora. The ones derived from the Spanish National Plan for Language Technologies are proficient solving several tasks and have been trained using large scale clean corpora. However, the Spanish Legal domain language could be think of an independent language on its own. We therefore created a Spanish Legal model from scratch trained exclusively on legal corpora.", "## Citing \n\n\nFor more information visit our GitHub repository", "## Funding\nThis work was funded by the Spanish State Secretariat for Digitalization and Artificial Intelligence (SEDIA) within the framework of the Plan-TL." ]
[ "TAGS\n#transformers #pytorch #roberta #fill-mask #legal #spanish #es #dataset-legal_ES #dataset-temu_legal #arxiv-2110.12201 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Spanish Legal-domain RoBERTa\n\nThere are few models trained for the Spanish language. Some of the models have been trained with a low resource, unclean corpora. The ones derived from the Spanish National Plan for Language Technologies are proficient solving several tasks and have been trained using large scale clean corpora. However, the Spanish Legal domain language could be think of an independent language on its own. We therefore created a Spanish Legal model from scratch trained exclusively on legal corpora.", "## Citing \n\n\nFor more information visit our GitHub repository", "## Funding\nThis work was funded by the Spanish State Secretariat for Digitalization and Artificial Intelligence (SEDIA) within the framework of the Plan-TL." ]
[ 76, 105, 14, 33 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #fill-mask #legal #spanish #es #dataset-legal_ES #dataset-temu_legal #arxiv-2110.12201 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Spanish Legal-domain RoBERTa\n\nThere are few models trained for the Spanish language. Some of the models have been trained with a low resource, unclean corpora. The ones derived from the Spanish National Plan for Language Technologies are proficient solving several tasks and have been trained using large scale clean corpora. However, the Spanish Legal domain language could be think of an independent language on its own. We therefore created a Spanish Legal model from scratch trained exclusively on legal corpora.## Citing \n\n\nFor more information visit our GitHub repository## Funding\nThis work was funded by the Spanish State Secretariat for Digitalization and Artificial Intelligence (SEDIA) within the framework of the Plan-TL." ]
[ -0.049268729984760284, 0.09738601744174957, -0.0030860528349876404, 0.04211772605776787, 0.0532393604516983, -0.027412232011556625, 0.15305471420288086, 0.05589287355542183, 0.05479232221841812, -0.022738821804523468, 0.04549996554851532, 0.04499076306819916, 0.012324036099016666, 0.09376964718103409, 0.026287483051419258, -0.19531168043613434, 0.029144613072276115, -0.06769857555627823, -0.11110582202672958, 0.042929261922836304, 0.10546594113111496, -0.03523019701242447, 0.08666052669286728, -0.01832231692969799, -0.043757617473602295, 0.02486739680171013, -0.029074665158987045, -0.09967250376939774, 0.0887889638543129, 0.09574590623378754, 0.1351124495267868, 0.03048398345708847, 0.03564101830124855, -0.012395882979035378, 0.015409181825816631, -0.011021650396287441, -0.031910210847854614, 0.027704430744051933, 0.026674533262848854, -0.046692855656147, 0.17084740102291107, -0.09374453872442245, 0.01720363460481167, -0.02473616786301136, -0.15417622029781342, 0.027189387008547783, -0.02404053509235382, -0.10260726511478424, 0.03031030483543873, 0.1115403100848198, 0.010561668314039707, 0.030692683532834053, -0.04547043889760971, 0.024058464914560318, -0.02309371903538704, -0.11196064949035645, -0.06119656190276146, 0.08117638528347015, 0.03994894027709961, 0.05311797931790352, 0.11471407115459442, 0.047465767711400986, 0.036564119160175323, 0.011958954855799675, 0.021969540044665337, -0.08270136266946793, -0.14116619527339935, -0.015050001442432404, -0.1652803272008896, -0.09439587593078613, 0.24675297737121582, -0.0979965552687645, -0.03882669284939766, -0.021275833249092102, -0.04499145969748497, 0.14021456241607666, 0.064118891954422, 0.013431782834231853, 0.00793434213846922, 0.02080564573407173, 0.10728765279054642, -0.07621777057647705, -0.052594926208257675, 0.004185875412076712, -0.1684974581003189, 0.23105087876319885, 0.024773649871349335, 0.020414475351572037, -0.05927813798189163, 0.024372104555368423, -0.0068522063083946705, -0.03951273113489151, 0.09869617223739624, 0.014720635488629341, -0.005182546563446522, 0.05379733815789223, -0.090905100107193, -0.1402835100889206, 0.08420741558074951, 0.08386670798063278, -0.022822106257081032, -0.07662340998649597, 0.03246983140707016, 0.10734199732542038, 0.039434634149074554, 0.12263105809688568, -0.09679283201694489, 0.07925484329462051, 0.05275501683354378, -0.10972509533166885, 0.027651231735944748, -0.04805832356214523, -0.19758504629135132, 0.029669681563973427, -0.1394200623035431, 0.07501602917909622, 0.05649638921022415, 0.06584595888853073, -0.0923195406794548, -0.03869415819644928, -0.04806821793317795, -0.06239917501807213, -0.026299597695469856, -0.09289253503084183, -0.010610890574753284, -0.08207545429468155, -0.04896698147058487, 0.016629021614789963, -0.0801357626914978, -0.048977140337228775, -0.003543374827131629, -0.08435676246881485, -0.1055426150560379, -0.08529017865657806, 0.0635889396071434, -0.09348532557487488, 0.06129790097475052, -0.16479328274726868, -0.17342902719974518, -0.08719532191753387, 0.08741555362939835, -0.10596862435340881, 0.015818921849131584, -0.10529676824808121, -0.017855122685432434, -0.07857519388198853, -0.06568248569965363, -0.07414740324020386, -0.02461603842675686, 0.049671608954668045, -0.09279464930295944, 0.04489573836326599, -0.2688051760196686, 0.01090976595878601, -0.14605455100536346, 0.008739250712096691, -0.036133941262960434, 0.12460945546627045, -0.03193386271595955, -0.02987809106707573, -0.08163128048181534, 0.03028494492173195, -0.011522513814270496, 0.1481035351753235, 0.011339856311678886, 0.09986225515604019, -0.14972282946109772, -0.030628837645053864, 0.10354040563106537, -0.03179606795310974, -0.029664264991879463, 0.143367201089859, -0.008896161802113056, 0.21926745772361755, 0.06295716017484665, 0.2265843152999878, 0.08926358073949814, -0.011705206707119942, 0.07577266544103622, -0.05162453278899193, -0.02541079930961132, -0.020211320370435715, 0.09508613497018814, 0.003847772255539894, -0.11478119343519211, 0.002165793674066663, -0.17615360021591187, 0.03099239245057106, -0.02750401943922043, -0.0807601734995842, 0.06351792812347412, -0.0060334037989377975, -0.012725473381578922, -0.04610135406255722, 0.06069201976060867, 0.011876976117491722, -0.002974174451082945, -0.041151441633701324, -0.015720129013061523, 0.0016695245867595077, -0.03163931891322136, -0.1564294993877411, 0.08444815874099731, 0.013743497431278229, 0.019022099673748016, -0.16076982021331787, -0.09361542016267776, 0.019039547070860863, 0.017123427242040634, 0.18064290285110474, 0.057438403367996216, -0.026677768677473068, -0.03639693930745125, -0.010167048312723637, 0.05236327275633812, -0.011046943254768848, -0.018601348623633385, 0.03261398896574974, -0.13900528848171234, 0.05887236073613167, -0.03211628645658493, -0.028691299259662628, 0.08597502112388611, 0.02665208838880062, -0.01574193499982357, 0.029585544019937515, -0.013458022847771645, 0.06351989507675171, -0.0009427699842490256, 0.14385725557804108, -0.03387980908155441, 0.008897983469069004, 0.07100227475166321, 0.004391408525407314, -0.14548829197883606, 0.05112027749419212, 0.0018674175953492522, 0.09154510498046875, 0.07062184065580368, -0.04000360518693924, 0.02331092581152916, -0.024284975603222847, 0.006322193890810013, 0.03382326290011406, -0.04743202403187752, -0.027292480692267418, 0.2856748402118683, -0.06955476105213165, 0.02774612233042717, -0.13138623535633087, 0.016320955008268356, 0.026152228936553, -0.059193000197410583, -0.06351260840892792, 0.07140336185693741, 0.16001665592193604, -0.017593426629900932, 0.12224853038787842, 0.12120252102613449, -0.0931931734085083, 0.19447462260723114, -0.021944688633084297, -0.013573768548667431, -0.016494836658239365, 0.031934287399053574, -0.0016818115254864097, 0.07695066183805466, -0.13636018335819244, -0.012955404818058014, 0.012680060230195522, -0.016528571024537086, 0.04839181527495384, -0.12262891978025436, -0.04702509194612503, -0.0335359200835228, -0.04401373490691185, -0.03709665685892105, 0.021320058032870293, -0.06119329482316971, 0.07326791435480118, 0.030753184109926224, -0.1376614272594452, 0.0613526776432991, -0.03103981539607048, -0.05355522409081459, 0.11679446697235107, -0.10120619088411331, -0.3393465578556061, -0.1333804875612259, 0.13355310261249542, -0.0396541990339756, 0.08310665190219879, 0.044232092797756195, -0.02768239751458168, -0.024949733167886734, 0.005347095895558596, 0.04516784846782684, 0.029955590143799782, -0.05274331197142601, -0.020661648362874985, 0.041780177503824234, -0.07423467934131622, -0.04600580781698227, -0.04834583401679993, 0.02430162951350212, -0.043383847922086716, 0.06828442215919495, -0.12393765896558762, 0.13806653022766113, 0.10237058997154236, -0.03552345559000969, -0.012403392232954502, -0.08054875582456589, 0.15129899978637695, -0.09287584573030472, -0.01973874494433403, 0.16117814183235168, -0.02020694501698017, -0.007401345297694206, 0.18658486008644104, 0.02515631914138794, -0.1044207215309143, -0.011175754480063915, -0.06656703352928162, -0.0904354453086853, -0.1873866468667984, -0.20587007701396942, -0.05716003105044365, -0.006964663974940777, 0.01798447221517563, 0.01726546697318554, -0.0298352912068367, 0.030557869002223015, 0.004075277596712112, 0.04005630686879158, 0.026945337653160095, 0.06136797368526459, 0.1585044413805008, -0.03110865131020546, 0.0318525992333889, -0.00989625696092844, -0.1349639743566513, 0.08821560442447662, 0.1415809839963913, 0.1846107840538025, 0.08943004906177521, 0.12124374508857727, 0.07345330715179443, 0.07014626264572144, -0.02433841861784458, -0.013568020425736904, 0.03359406441450119, 0.08452928066253662, -0.13163688778877258, -0.0544612891972065, -0.05822060629725456, 0.04365706071257591, -0.028758563101291656, -0.1194615513086319, -0.024283023551106453, -0.00951728131622076, 0.06556756794452667, 0.10327102988958359, -0.05433857813477516, -0.1796213835477829, -0.032861191779375076, 0.05317327752709389, 0.02860759198665619, -0.09372137486934662, 0.0794726237654686, 0.11914375424385071, -0.08263539522886276, 0.008004480041563511, 0.010308537632226944, 0.0842597559094429, -0.05650246515870094, 0.03290243074297905, -0.07248218357563019, -0.015854137018322945, 0.07131628692150116, 0.09551183134317398, -0.2787322402000427, 0.2789829671382904, 0.028955003246665, 0.016836494207382202, -0.15830041468143463, -0.02371067926287651, -0.06753429025411606, 0.03622075170278549, 0.1570461243391037, 0.043150484561920166, -0.049886077642440796, 0.012581118382513523, -0.03552506864070892, 0.06832918524742126, 0.05087053403258324, -0.037477631121873856, 0.028808310627937317, -0.009991632774472237, -0.004810419399291277, -0.0021185653749853373, -0.1324908286333084, 0.07544078677892685, -0.15127015113830566, 0.044273603707551956, 0.06479877978563309, 0.09175563603639603, -0.033180683851242065, -0.03149940446019173, 0.04988505691289902, 0.10414673388004303, -0.07936636358499527, -0.08511039614677429, -0.07087880373001099, -0.026782777160406113, 0.08468621224164963, -0.005688345525413752, 0.04612903669476509, 0.03111095540225506, -0.009265735745429993, -0.03769359365105629, -0.04596152529120445, 0.14796262979507446, -0.14517873525619507, -0.08640103787183762, -0.08907601237297058, -0.0075830151326954365, 0.016304343938827515, 0.030429068952798843, 0.008169217966496944, -0.015107369050383568, -0.029316751286387444, -0.03890477120876312, -0.04361201450228691, 0.11023752391338348, 0.018781328573822975, 0.12268117070198059, -0.07708039879798889, -0.09327489137649536, -0.037269119173288345, -0.13116781413555145, 0.037081386893987656, 0.08765240758657455, -0.011966961435973644, 0.12583661079406738, 0.21538497507572174, -0.14652115106582642, -0.2692711651325226, -0.05972619354724884, -0.010641269385814667, 0.0487946979701519, 0.006476485636085272, -0.19393236935138702, -0.09599649161100388, 0.08094798773527145, -0.016627175733447075, -0.14748603105545044, -0.30990591645240784, -0.07847051322460175, 0.07862304151058197, 0.012270674109458923, 0.2219248265028, -0.10959426313638687, -0.059005994349718094, -0.051015958189964294, 0.16800770163536072, 0.12682823836803436, -0.08505138009786606, 0.04667111113667488, 0.05057203397154808, 0.04347207769751549, 0.02546466700732708, 0.015495045110583305, 0.09802363812923431, -0.04876997694373131, 0.04154425859451294, -0.007613197434693575, -0.05147958919405937, 0.22256071865558624, 0.05492045357823372, 0.09317399561405182, 0.09599477797746658, 0.03228523209691048, -0.0024184260983020067, -0.05896714702248573, -0.07331305742263794, 0.03615560382604599, 0.008504373952746391, -0.10370194166898727, -0.09257974475622177, 0.09470494091510773, 0.009501011110842228, 0.000862341548781842, 0.1181337758898735, -0.020319655537605286, -0.06651079654693604, 0.027594175189733505, 0.22307799756526947, -0.0005041243857704103, 0.10133194178342819, 0.05195598676800728, -0.0418209508061409, 0.07842452824115753, -0.038307469338178635, -0.027798952534794807, 0.12420300394296646, -0.056712958961725235, 0.06179332733154297, 0.016318924725055695, -0.08431559801101685, 0.026815956458449364, 0.13761481642723083, -0.034317828714847565, -0.20307065546512604, -0.033770497888326645, 0.10883114486932755, 0.050704557448625565, 0.05854098126292229, 0.10095503181219101, -0.049766022711992264, -0.05304785817861557, 0.005327779799699783, 0.004651150666177273, -0.013884796760976315, 0.025741172954440117, 0.00871748011559248, -0.04783684015274048, -0.06508944183588028, 0.1257716417312622, 0.10286729782819748, -0.14620184898376465, 0.08935188502073288, -0.03217672184109688, -0.1003079041838646, -0.044372156262397766, 0.03596953675150871, 0.009970092214643955, -0.25767213106155396, -0.12043748795986176, -0.14034521579742432, -0.11062001436948776, -0.0023512723855674267, 0.1519167721271515, 0.029846929013729095, -0.024694958701729774, -0.000018352689949097112, 0.019639186561107635, 0.02479473315179348, 0.028110994026064873, -0.021291080862283707, -0.039705660194158554, -0.037854209542274475, 0.01395262498408556, 0.0013935415772721171, 0.007112229708582163, -0.06743434071540833, -0.04354848712682724, -0.19179116189479828, 0.07065919041633606, -0.09636800736188889, 0.03962637111544609, -0.06730800867080688, -0.010411296039819717, 0.04102558270096779, -0.04591402783989906, 0.0013289520284160972, -0.009063015691936016, -0.07071945071220398, 0.047209374606609344, -0.014045175164937973, 0.058803316205739975, 0.0162754338234663, -0.04924993962049484, 0.0023003672249615192, 0.035773832350969315, 0.04752327874302864, -0.052521221339702606, -0.03684268891811371, 0.06996987015008926, -0.22011925280094147, 0.07768616825342178, 0.09517268091440201, 0.06996132433414459, 0.028745997697114944, -0.055775199085474014, 0.0019009871175512671, 0.09230107069015503, -0.016062602400779724, 0.018333951011300087, 0.019082291051745415, -0.04883453994989395, 0.11620204150676727, 0.12816400825977325, -0.10173070430755615, -0.016380924731492996, 0.04339359700679779, 0.06966116279363632, 0.01721019670367241, 0.10900222510099411, -0.05473128333687782, -0.006355507764965296, 0.04313594847917557, 0.0276838056743145, -0.02590794488787651, -0.051314543932676315, -0.14485813677310944, -0.14400887489318848, 0.01751386560499668, 0.009659972973167896, 0.26806148886680603, 0.174974724650383, 0.026216527447104454, -0.026049014180898666, -0.0013687763130292296, 0.029860179871320724, 0.0300822164863348, 0.014467642642557621, -0.009919223375618458, 0.046790674328804016, -0.16460610926151276, 0.019078154116868973, -0.04742760583758354, 0.07670832425355911, 0.042541779577732086, 0.020344218239188194, 0.15770287811756134, 0.07865066081285477, 0.12044721841812134, 0.04847494885325432, -0.09381429105997086, -0.07778678089380264, 0.10935895889997482, 0.043308183550834656, -0.06928497552871704, 0.04727507010102272, 0.10204862803220749, -0.08872511982917786, 0.10258863866329193, 0.025629255920648575, -0.004350694827735424, -0.09914353489875793, -0.26018404960632324, -0.0125456303358078, -0.061976343393325806, -0.05995681881904602, -0.11386547237634659, 0.02781181037425995, -0.027658484876155853, 0.0044496371410787106, -0.08253876119852066, 0.11490463465452194, -0.0808618888258934, -0.09674938023090363, 0.04199324548244476, -0.0016966296825557947, 0.16700585186481476, -0.07310453057289124, -0.049390025436878204, -0.09282752126455307, 0.11657186597585678, -0.0380760133266449, 0.08519947528839111, 0.04099305719137192, -0.0614653043448925, -0.058453626930713654, -0.009251594543457031, -0.06323352456092834, 0.029901644214987755, 0.006638174410909414, 0.11316021531820297, 0.02399294450879097, -0.1087985709309578, 0.09294958412647247, 0.17660656571388245, -0.029928913339972496, -0.08748907595872879, -0.11129171401262283, 0.23485814034938812, -0.04194771125912666, 0.03447018191218376, -0.0035480123478919268, -0.03828883916139603, -0.019713839516043663, 0.21209962666034698, 0.26860830187797546, -0.01456111017614603, -0.02288656122982502, 0.002764855045825243, 0.01534189097583294, -0.001207670196890831, 0.15697450935840607, -0.023981692269444466, 0.3501249849796295, -0.05390891432762146, -0.030595796182751656, -0.039651066064834595, 0.05661948397755623, -0.09498178958892822, -0.010252650827169418, -0.04058264568448067, -0.11152608692646027, -0.030743269249796867, 0.1228952631354332, -0.1591438502073288, -0.1883286088705063, 0.04971073940396309, -0.09038836508989334, -0.059993572533130646, -0.02416553534567356, -0.05678700655698776, 0.034866414964199066, 0.09680373221635818, -0.03264767676591873, -0.025414690375328064, -0.009786217473447323, 0.044618383049964905, -0.11559142917394638, 0.0054155676625669, 0.06862562149763107, 0.052996039390563965, 0.1643151342868805, -0.0068765017203986645, 0.1275274008512497, 0.05247765779495239, 0.019571978598833084, -0.08126168698072433, 0.09005644917488098, 0.05145176872611046, 0.07569711655378342, 0.0667494460940361, -0.10308106988668442, -0.031752366572618484, 0.1400645524263382, 0.09229708462953568, 0.013364340178668499, 0.07070861011743546, 0.08602204918861389, 0.0010496375616639853, -0.07272687554359436, 0.07354152202606201, -0.16363327205181122, 0.07586735486984253, 0.12885668873786926, 0.011594844050705433, 0.0035467660054564476, -0.040030162781476974, 0.057913780212402344, 0.046631522476673126, 0.1341187208890915, -0.026644861325621605, -0.07967589795589447, -0.056702546775341034, 0.04813166707754135, 0.013454476371407509, -0.1601286679506302, 0.02788538672029972, -0.07106612622737885, 0.031441573053598404, -0.04513641819357872, 0.06254331767559052, 0.06926599144935608, -0.012747020460665226, -0.00461618835106492, -0.01807662472128868, -0.01845952868461609, 0.026797715574502945, -0.08445671945810318, -0.0011423079995438457 ]
null
null
transformers
# GPT2-large trained with data from National Library of Spain (BNE) ## Model Description GPT2-large-bne is a transformer-based model for the Spanish language. It is based on the [GPT-2](http://www.persagen.com/files/misc/radford2019language.pdf) model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019. ## Training corpora and preprocessing The [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) crawls all .es domains once a year. The training corpus consists of 59TB of WARC files from these crawls, carried out from 2009 to 2019. To obtain a high-quality training corpus, the corpus has been preprocessed with a pipeline of operations, including among the others, sentence splitting, language detection, filtering of bad-formed sentences and deduplication of repetitive contents. During the process document boundaries are kept. This resulted into 2TB of Spanish clean corpus. Further global deduplication among the corpus is applied, resulting into 570GB of text. Some of the statistics of the corpus: | Corpora | Number of documents | Number of tokens | Size (GB) | |---------|---------------------|------------------|-----------| | BNE | 201,080,084 | 135,733,450,668 | 570GB | ## Tokenization and pre-training The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original [GPT-2](http://www.persagen.com/files/misc/radford2019language.pdf) model with a vocabulary size of 50,262 tokens. The GPT2-large-bne pre-training consists of an autoregressive language model training that follows the approach of the GPT-2. The training lasted a total of 10 days with 32 computing nodes each one with 4 NVIDIA V100 GPUs of 16GB VRAM. ## Evaluation and results For evaluation details visit our [GitHub repository](https://github.com/PlanTL-SANIDAD/lm-spanish). ## Citing Check out our paper for all the details: https://arxiv.org/abs/2107.07253 ``` @misc{gutierrezfandino2021spanish, title={Spanish Language Models}, author={Asier Gutiérrez-Fandiño and Jordi Armengol-Estapé and Marc Pàmies and Joan Llop-Palao and Joaquín Silveira-Ocampo and Casimiro Pio Carrino and Aitor Gonzalez-Agirre and Carme Armentano-Oller and Carlos Rodriguez-Penagos and Marta Villegas}, year={2021}, eprint={2107.07253}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
{"language": ["es"], "license": "apache-2.0", "tags": ["national library of spain", "spanish", "bne"], "datasets": ["bne"], "metrics": ["ppl"]}
text-generation
BSC-LT/gpt2-large-bne
[ "transformers", "pytorch", "gpt2", "text-generation", "national library of spain", "spanish", "bne", "es", "dataset:bne", "arxiv:2107.07253", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[ "2107.07253" ]
[ "es" ]
TAGS #transformers #pytorch #gpt2 #text-generation #national library of spain #spanish #bne #es #dataset-bne #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
GPT2-large trained with data from National Library of Spain (BNE) ================================================================= Model Description ----------------- GPT2-large-bne is a transformer-based model for the Spanish language. It is based on the GPT-2 model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019. Training corpora and preprocessing ---------------------------------- The National Library of Spain (Biblioteca Nacional de España) crawls all .es domains once a year. The training corpus consists of 59TB of WARC files from these crawls, carried out from 2009 to 2019. To obtain a high-quality training corpus, the corpus has been preprocessed with a pipeline of operations, including among the others, sentence splitting, language detection, filtering of bad-formed sentences and deduplication of repetitive contents. During the process document boundaries are kept. This resulted into 2TB of Spanish clean corpus. Further global deduplication among the corpus is applied, resulting into 570GB of text. Some of the statistics of the corpus: Tokenization and pre-training ----------------------------- The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original GPT-2 model with a vocabulary size of 50,262 tokens. The GPT2-large-bne pre-training consists of an autoregressive language model training that follows the approach of the GPT-2. The training lasted a total of 10 days with 32 computing nodes each one with 4 NVIDIA V100 GPUs of 16GB VRAM. Evaluation and results ---------------------- For evaluation details visit our GitHub repository. Citing ------ Check out our paper for all the details: URL
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #national library of spain #spanish #bne #es #dataset-bne #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 85 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #national library of spain #spanish #bne #es #dataset-bne #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0801360160112381, 0.1883486658334732, -0.007008853834122419, 0.08262328058481216, 0.06631986796855927, 0.024362949654459953, 0.08099566400051117, 0.13533808290958405, -0.018353717401623726, -0.01270813588052988, 0.14412915706634521, 0.21073907613754272, 0.014468490146100521, -0.024861140176653862, -0.07362896203994751, -0.17949305474758148, 0.06223294511437416, 0.046173758804798126, -0.05569081753492355, 0.05282722786068916, 0.09522499889135361, -0.016760483384132385, 0.059611983597278595, -0.04392340034246445, -0.07399705052375793, 0.05841990187764168, 0.010456441901624203, -0.1586655229330063, 0.13670594990253448, 0.08977686613798141, 0.055286552757024765, 0.045737143605947495, -0.03173968195915222, -0.10716374218463898, -0.005158239975571632, -0.036550067365169525, -0.09872878342866898, 0.08626864850521088, 0.08088789880275726, -0.07659375667572021, 0.11397528648376465, 0.03871380537748337, -0.05325987562537193, 0.02011999860405922, -0.16623002290725708, -0.18402382731437683, -0.10735709220170975, 0.007274860516190529, -0.013964453712105751, 0.06826270371675491, 0.029977696016430855, 0.07131842523813248, -0.04577649012207985, 0.014429246075451374, 0.14505670964717865, -0.2993553578853607, -0.018615180626511574, -0.007412807550281286, 0.07038306444883347, 0.08856339752674103, 0.008013788610696793, 0.07268533855676651, 0.10571368038654327, -0.00840675737708807, 0.0014095681253820658, -0.10066898167133331, -0.1738399863243103, 0.0503433421254158, -0.04186485707759857, -0.08445364981889725, 0.27993690967559814, -0.054772667586803436, 0.043757613748311996, 0.029512852430343628, -0.07042919844388962, 0.07899924367666245, 0.03214700147509575, 0.02607615292072296, 0.0018287193961441517, 0.05250868573784828, 0.1258673071861267, -0.04412733390927315, -0.09939103573560715, -0.005373053252696991, -0.19027313590049744, 0.09572108089923859, 0.009612346068024635, 0.048430874943733215, -0.034662164747714996, 0.07064312696456909, 0.010747183114290237, -0.12354640662670135, 0.01768854446709156, -0.033484406769275665, 0.13727495074272156, 0.06359154731035233, -0.030211981385946274, -0.008557259105145931, 0.12053682655096054, 0.14065250754356384, -0.0704449862241745, -0.04579957202076912, -0.07890215516090393, 0.12002818286418915, 0.01813753880560398, 0.05018303915858269, -0.012675585225224495, -0.06563540548086166, 0.06623955816030502, -0.11404973268508911, 0.049643516540527344, -0.01062091812491417, -0.18949776887893677, -0.09050863981246948, -0.027339832857251167, 0.11399716138839722, 0.08361227810382843, -0.02283306047320366, -0.049665533006191254, -0.007219797931611538, 0.12542685866355896, -0.037117086350917816, 0.017932897433638573, -0.01773436926305294, -0.03213179484009743, 0.04347042366862297, -0.024465864524245262, -0.017794985324144363, -0.10289610922336578, 0.003167712362483144, -0.07178660482168198, -0.0717000961303711, -0.0411960631608963, -0.04590882733464241, 0.10520277172327042, -0.10095053166151047, 0.05878470838069916, -0.1612791121006012, -0.19303929805755615, 0.0070110191591084, 0.0345318540930748, -0.11466508358716965, -0.0363287478685379, -0.034458283334970474, 0.015924587845802307, 0.05832803249359131, -0.09390780329704285, 0.04278865084052086, -0.10641386359930038, 0.1242510974407196, -0.02567378617823124, 0.035062965005636215, -0.21600870788097382, 0.018350599333643913, -0.11694281548261642, -0.007582386024296284, -0.10194206982851028, 0.03739161789417267, -0.061698827892541885, 0.04762658104300499, -0.07358500361442566, -0.0027206242084503174, -0.004969422239810228, 0.055419858545064926, 0.013727592304348946, 0.13281027972698212, -0.07695610076189041, -0.08844634890556335, 0.18247056007385254, -0.06050338223576546, -0.150785431265831, 0.0904870331287384, 0.03015405684709549, 0.07655806839466095, 0.07552004605531693, 0.1995723843574524, -0.014868093654513359, -0.019752593711018562, 0.006392308976501226, 0.07803123444318771, -0.013686513528227806, -0.1881076991558075, 0.09498041868209839, -0.04960063844919205, -0.06938323378562927, 0.04030468314886093, -0.05094198137521744, 0.05119138956069946, -0.005938315764069557, -0.035746846348047256, 0.0009184439550153911, -0.016342591494321823, 0.028347719460725784, -0.003352118656039238, 0.07787653803825378, -0.06552137434482574, -0.010939106345176697, 0.03912598267197609, -0.025390956550836563, 0.031264692544937134, 0.0384753942489624, -0.006500889081507921, 0.14339803159236908, -0.052394021302461624, 0.039299316704273224, -0.10039641708135605, 0.0675467923283577, -0.011156772263348103, 0.08747486025094986, 0.022970926016569138, 0.008882745169103146, 0.01602134481072426, -0.01961853913962841, -0.059183742851018906, 0.005875087343156338, 0.06913632154464722, -0.013395027257502079, -0.014413785189390182, -0.1471678763628006, 0.10283463448286057, -0.025957273319363594, -0.03683415800333023, -0.06817124783992767, 0.008143489249050617, 0.00683237798511982, 0.05193668603897095, -0.07610864192247391, 0.0918339267373085, -0.061961185187101364, 0.05318932980298996, -0.08791027218103409, -0.004252101294696331, 0.0860445648431778, 0.04141540825366974, -0.06416085362434387, 0.16203366219997406, -0.07820520550012589, 0.23517848551273346, 0.1760854572057724, -0.17413805425167084, 0.06717200577259064, -0.048975955694913864, -0.003902419237419963, 0.015472334809601307, 0.023366162553429604, -0.06054282560944557, 0.08998509496450424, -0.005376207176595926, 0.16527105867862701, -0.14885255694389343, -0.04140903800725937, 0.016279831528663635, -0.05215451866388321, -0.043775275349617004, 0.1341959536075592, 0.14985527098178864, -0.09822019189596176, 0.20650503039360046, 0.2186538130044937, -0.021448375657200813, 0.21060407161712646, -0.009713983163237572, -0.03943614289164543, 0.043959274888038635, -0.04102311655879021, 0.018436510115861893, 0.09239956736564636, -0.13586489856243134, 0.0378698855638504, 0.08333545923233032, 0.033551447093486786, 0.07530524581670761, -0.1338377147912979, -0.07111085206270218, -0.026564158499240875, -0.038742102682590485, -0.0088766198605299, 0.039694592356681824, -0.020813435316085815, 0.14144983887672424, 0.008679867722094059, -0.09004849195480347, 0.07170390337705612, 0.028324777260422707, -0.07225041091442108, 0.1722748875617981, -0.15961246192455292, -0.3126392066478729, -0.18504935503005981, -0.06993624567985535, -0.01619877479970455, 0.08599768579006195, 0.12319041788578033, -0.061741817742586136, -0.005812083836644888, 0.04051990807056427, 0.07175686955451965, -0.03858856484293938, -0.07228697836399078, -0.04250870645046234, 0.06677750498056412, -0.06797367334365845, -0.1433720588684082, -0.030805109068751335, 0.04425555467605591, -0.03288807347416878, 0.05062379688024521, -0.11291106045246124, 0.12313380092382431, 0.08372380584478378, 0.050300855189561844, -0.0015603682259097695, -0.04778250679373741, 0.1187148466706276, -0.06983562558889389, -0.008933890610933304, 0.1848328858613968, 0.021212754771113396, 0.019682038575410843, 0.11760137230157852, 0.03246012330055237, -0.07053713500499725, -0.04922803118824959, -0.04655010253190994, -0.06927317380905151, -0.3143453598022461, -0.13481304049491882, -0.08815427124500275, 0.07774548977613449, 0.060521312057971954, 0.05375967547297478, 0.1426703780889511, 0.0922267884016037, -0.020815322175621986, 0.03680559992790222, -0.028582531958818436, 0.08253544569015503, 0.26208004355430603, -0.014087839983403683, 0.08272980153560638, -0.09003140032291412, -0.06596602499485016, 0.14490848779678345, 0.14470024406909943, 0.06313539296388626, 0.07496441155672073, 0.13746461272239685, 0.026677008718252182, 0.10337910056114197, 0.016985351219773293, 0.09849465638399124, 0.020201191306114197, 0.005420507863163948, -0.09385248273611069, -0.04062953591346741, -0.055423200130462646, -0.00808027759194374, -0.08614028990268707, -0.05800805613398552, -0.04885130375623703, -0.21861796081066132, 0.08993730694055557, 0.05114053189754486, 0.019120344892144203, -0.17910698056221008, 0.020305855199694633, 0.07032109051942825, 0.0010885135270655155, -0.13505013287067413, 0.04310744255781174, -0.02790042571723461, -0.11538165807723999, 0.07227066904306412, 0.010316770523786545, 0.1408987194299698, -0.04944781959056854, 0.048422425985336304, -0.12704937160015106, -0.10895822197198868, 0.029283346608281136, 0.1273130476474762, -0.2645379602909088, 0.2620027959346771, 0.024988558143377304, -0.019833361729979515, -0.06498336046934128, 0.01110709086060524, -0.02193514071404934, 0.11494315415620804, 0.11004411429166794, 0.006102091632783413, 0.046609777957201004, 0.008205943740904331, -0.04684508219361305, 0.05278811231255531, -0.05866837128996849, -0.05332259461283684, -0.023065689951181412, -0.007054266054183245, 0.02085029147565365, -0.024460719898343086, 0.018085209652781487, 0.0007262082654051483, -0.1801890879869461, 0.0458328016102314, 0.01586666889488697, 0.07016945630311966, -0.011063278652727604, -0.041982557624578476, -0.10482610017061234, 0.15807504951953888, -0.12121981382369995, -0.09492087364196777, -0.0902874693274498, -0.03313520923256874, 0.09492333978414536, -0.022432982921600342, 0.018492694944143295, -0.03628162294626236, -0.026059914380311966, -0.06921061873435974, -0.11844867467880249, 0.11126291006803513, -0.1239190474152565, -0.027446815744042397, -0.08187134563922882, 0.11340693384408951, -0.06017119064927101, 0.032007697969675064, 0.01765996590256691, 0.007298867218196392, -0.062384381890296936, -0.07831459492444992, 0.008515519089996815, -0.04392218217253685, 0.10345881432294846, -0.02833503484725952, -0.11289720982313156, -0.04318559169769287, 0.0012872848892584443, -0.11332736164331436, 0.1688283085823059, 0.18591715395450592, -0.03405104577541351, 0.14216947555541992, 0.23932115733623505, -0.15602360665798187, -0.22022731602191925, -0.13414797186851501, -0.09647081047296524, -0.035571467131376266, -0.09695884585380554, -0.22841452062129974, 0.05149365961551666, 0.1898987591266632, -0.05586102232336998, 0.09772440046072006, -0.30193987488746643, -0.05622134730219841, 0.09584055840969086, -0.019035272300243378, 0.3401777744293213, -0.1465398669242859, -0.12007799744606018, -0.10566786676645279, -0.1267300546169281, 0.18366354703903198, -0.06972172111272812, 0.053525082767009735, -0.04109669476747513, -0.014688397757709026, -0.002127552404999733, 0.012579438276588917, 0.16411828994750977, -0.056495875120162964, -0.02262173593044281, -0.061981212347745895, 0.028633825480937958, 0.13120214641094208, 0.014861706644296646, -0.0023629150819033384, -0.10190176963806152, -0.016687843948602676, -0.1623268872499466, -0.027003074064850807, -0.05582491308450699, 0.0765751525759697, -0.0060523259453475475, -0.04340876638889313, -0.016934648156166077, -0.04303564503788948, -0.015117794275283813, -0.008951563388109207, 0.1847257763147354, 0.019354112446308136, 0.09827007353305817, 0.1327301561832428, 0.11285988986492157, -0.10413976013660431, 0.03828483447432518, -0.06797029823064804, -0.045212551951408386, 0.07975256443023682, -0.08336610347032547, 0.004671028349548578, 0.13740131258964539, -0.05716332420706749, 0.02993115969002247, 0.046959288418293, -0.04222416505217552, -0.014394677244126797, 0.15147125720977783, -0.1307528167963028, 0.03782421723008156, -0.02074291557073593, 0.04491487145423889, 0.10172601789236069, 0.008428438566625118, 0.14395540952682495, 0.04214177653193474, -0.06795857101678848, 0.020812377333641052, -0.02279861643910408, -0.037517815828323364, 0.11003296077251434, 0.027659481391310692, -0.015123575925827026, -0.12224354594945908, 0.11096218228340149, 0.05083683505654335, -0.13623599708080292, 0.01722407527267933, 0.13466300070285797, -0.06128133460879326, -0.10453520715236664, 0.03751489520072937, 0.10544920712709427, -0.30468764901161194, -0.10514752566814423, -0.11293196678161621, -0.09910569339990616, 0.07749560475349426, 0.17503796517848969, 0.05055520683526993, 0.03976467624306679, 0.004017524886876345, -0.0165396835654974, 0.03610554337501526, -0.0035421543288975954, -0.06426837295293808, 0.025461887940764427, -0.04575628787279129, -0.04981369897723198, -0.0007132256869226694, 0.03283063322305679, -0.041480906307697296, 0.025241779163479805, -0.1561027318239212, 0.037833668291568756, -0.1666066199541092, 0.031539976596832275, -0.07227630913257599, -0.040037207305431366, -0.052895598113536835, -0.058369915932416916, -0.03246722370386124, -0.08088221400976181, -0.0911984071135521, -0.0005766332615166903, -0.017796609550714493, 0.10446907579898834, -0.08752219378948212, -0.02568414807319641, 0.05999015271663666, -0.008896913379430771, 0.09353361278772354, 0.06160961464047432, -0.022196205332875252, 0.1285657435655594, -0.18881793320178986, -0.006065935827791691, 0.0980796366930008, 0.0533602274954319, 0.04433094337582588, 0.03065890073776245, 0.020620372146368027, 0.11785656958818436, -0.009483888745307922, 0.0639297366142273, -0.04797174781560898, -0.14011548459529877, 0.017203591763973236, 0.025809423997998238, -0.1714201420545578, 0.019748441874980927, 0.01437630970031023, 0.14224538207054138, -0.035332027822732925, 0.07237511873245239, -0.06267418712377548, -0.004375973250716925, -0.047019172459840775, 0.03848719969391823, -0.003673041705042124, -0.12341520935297012, -0.1252204030752182, -0.03938977047801018, -0.011910752393305302, 0.03736802935600281, 0.284665048122406, 0.07253269851207733, -0.00470698531717062, 0.021011970937252045, 0.042159777134656906, 0.02416616678237915, 0.0030914791859686375, 0.25604814291000366, 0.08895595371723175, -0.00882903952151537, -0.14735311269760132, 0.08351153135299683, 0.031482305377721786, 0.02624538168311119, 0.028866074979305267, 0.10387217253446579, 0.1974313110113144, 0.07429879158735275, 0.039077628403902054, -0.03109002858400345, -0.01037408784031868, -0.04977156221866608, 0.09549514949321747, 0.05051907151937485, 0.03992842882871628, 0.13253426551818848, 0.15215766429901123, -0.053875118494033813, 0.036967940628528595, -0.007375017274171114, 0.007661676034331322, -0.11633691936731339, -0.1305561512708664, -0.03293874114751816, -0.12823769450187683, -0.026759618893265724, -0.10063888132572174, 0.06527300924062729, 0.0036043846048414707, 0.04989313706755638, -0.0940319150686264, -0.07748573273420334, 0.09414525330066681, -0.13724680244922638, 0.031306080520153046, 0.018567783758044243, 0.1126350685954094, -0.12589634954929352, 0.02854122966527939, -0.06388894468545914, 0.08510321378707886, -0.025181304663419724, 0.08938050270080566, 0.013458897359669209, 0.0035758723970502615, -0.09522747993469238, -0.028332900255918503, -0.02338240295648575, 0.062390197068452835, 0.00982937216758728, 0.1840154081583023, -0.007476424798369408, -0.0004376040888018906, 0.0804324820637703, 0.23115675151348114, -0.010176983661949635, -0.07235216349363327, -0.02520243264734745, 0.11181630194187164, 0.005881716497242451, 0.10901448130607605, -0.011996614746749401, -0.002105233259499073, -0.0545138344168663, 0.2581590414047241, 0.2872374951839447, -0.050606004893779755, -0.03280387818813324, 0.025969693437218666, 0.028375517576932907, 0.11222794651985168, 0.0733812227845192, 0.0987434834241867, 0.30511367321014404, -0.08626706153154373, -0.05918455123901367, -0.07972778379917145, 0.09949769079685211, -0.12473408877849579, 0.1255909502506256, -0.030505668371915817, -0.10669004172086716, -0.0014789410633966327, 0.10639352351427078, -0.1286339908838272, -0.018081437796354294, -0.037105344235897064, -0.15595410764217377, -0.09336380660533905, -0.004413299728184938, 0.11054351180791855, 0.036552537232637405, 0.03681106120347977, 0.004450411070138216, -0.06526514887809753, 0.0084838280454278, 0.010990566574037075, -0.20805838704109192, -0.09737144410610199, 0.045444056391716, -0.03002588078379631, 0.18255667388439178, -0.024101803079247475, 0.04379257187247276, 0.08916056901216507, 0.04598982259631157, -0.10091624408960342, 0.03307325392961502, 0.04605608433485031, -0.0009808276081457734, 0.09166257828474045, -0.12085720896720886, 0.03413432464003563, -0.026271983981132507, 0.07093841582536697, -0.030156973749399185, 0.048008065670728683, 0.11886686831712723, -0.06346733868122101, -0.03992347791790962, 0.06743384152650833, -0.10675883293151855, 0.021249540150165558, 0.05279889702796936, -0.008768281899392605, -0.017712483182549477, -0.06775373220443726, 0.002576753729954362, 0.051583923399448395, -0.01795348897576332, -0.038859352469444275, -0.11342667788267136, -0.05200405791401863, 0.09565962105989456, 0.0015090311644598842, -0.17730888724327087, 0.040915943682193756, -0.09990733861923218, 0.02451866865158081, -0.15793611109256744, 0.013162867166101933, 0.028429048135876656, 0.006413837429136038, 0.00631957221776247, -0.08127374947071075, -0.00734015041962266, 0.011015532538294792, -0.028950074687600136, -0.05002220720052719 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/roberta-base-biomedical-clinical-es # Biomedical-clinical language model for Spanish Biomedical pretrained language model for Spanish. For more details about the corpus, the pretraining and the evaluation, check the official [repository](https://github.com/PlanTL-SANIDAD/lm-biomedical-clinical-es) and read our [preprint](https://arxiv.org/abs/2109.03570) "_Carrino, C. P., Armengol-Estapé, J., Gutiérrez-Fandiño, A., Llop-Palao, J., Pàmies, M., Gonzalez-Agirre, A., & Villegas, M. (2021). Biomedical and Clinical Language Models for Spanish: On the Benefits of Domain-Specific Pretraining in a Mid-Resource Scenario._". ## Tokenization and model pretraining This model is a [RoBERTa-based](https://github.com/pytorch/fairseq/tree/master/examples/roberta) model trained on a **biomedical-clinical** corpus in Spanish collected from several sources (see next section). The training corpus has been tokenized using a byte version of [Byte-Pair Encoding (BPE)](https://github.com/openai/gpt-2) used in the original [RoBERTA](https://github.com/pytorch/fairseq/tree/master/examples/roberta) model with a vocabulary size of 52,000 tokens. The pretraining consists of a masked language model training at the subword level following the approach employed for the RoBERTa base model with the same hyperparameters as in the original work. The training lasted a total of 48 hours with 16 NVIDIA V100 GPUs of 16GB DDRAM, using Adam optimizer with a peak learning rate of 0.0005 and an effective batch size of 2,048 sentences. ## Training corpora and preprocessing The training corpus is composed of several biomedical corpora in Spanish, collected from publicly available corpora and crawlers, and a real-world clinical corpus collected from more than 278K clinical documents and notes. To obtain a high-quality training corpus while retaining the idiosyncrasies of the clinical language, a cleaning pipeline has been applied only to the biomedical corpora, keeping the clinical corpus uncleaned. Essentially, the cleaning operations used are: - data parsing in different formats - sentence splitting - language detection - filtering of ill-formed sentences - deduplication of repetitive contents - keep the original document boundaries Then, the biomedical corpora are concatenated and further global deduplication among the biomedical corpora have been applied. Eventually, the clinical corpus is concatenated to the cleaned biomedical corpus resulting in a medium-size biomedical-clinical corpus for Spanish composed of more than 1B tokens. The table below shows some basic statistics of the individual cleaned corpora: | Name | No. tokens | Description | |-----------------------------------------------------------------------------------------|-------------|------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | [Medical crawler](https://zenodo.org/record/4561970) | 745,705,946 | Crawler of more than 3,000 URLs belonging to Spanish biomedical and health domains. | | Clinical cases misc. | 102,855,267 | A miscellany of medical content, essentially clinical cases. Note that a clinical case report is a scientific publication where medical practitioners share patient cases and it is different from a clinical note or document. | | Clinical notes/documents | 91,250,080 | Collection of more than 278K clinical documents, including discharge reports, clinical course notes and X-ray reports, for a total of 91M tokens. | | [Scielo](https://github.com/PlanTL-SANIDAD/SciELO-Spain-Crawler) | 60,007,289 | Publications written in Spanish crawled from the Spanish SciELO server in 2017. | | [BARR2_background](https://temu.bsc.es/BARR2/downloads/background_set.raw_text.tar.bz2) | 24,516,442 | Biomedical Abbreviation Recognition and Resolution (BARR2) containing Spanish clinical case study sections from a variety of clinical disciplines. | | Wikipedia_life_sciences | 13,890,501 | Wikipedia articles crawled 04/01/2021 with the [Wikipedia API python library](https://pypi.org/project/Wikipedia-API/) starting from the "Ciencias\_de\_la\_vida" category up to a maximum of 5 subcategories. Multiple links to the same articles are then discarded to avoid repeating content. | | Patents | 13,463,387 | Google Patent in Medical Domain for Spain (Spanish). The accepted codes (Medical Domain) for Json files of patents are: "A61B", "A61C","A61F", "A61H", "A61K", "A61L","A61M", "A61B", "A61P". | | [EMEA](http://opus.nlpl.eu/download.php?f=EMEA/v3/moses/en-es.txt.zip) | 5,377,448 | Spanish-side documents extracted from parallel corpora made out of PDF documents from the European Medicines Agency. | | [mespen_Medline](https://zenodo.org/record/3562536#.YTt1fH2xXbR) | 4,166,077 | Spanish-side articles extracted from a collection of Spanish-English parallel corpus consisting of biomedical scientific literature. The collection of parallel resources are aggregated from the MedlinePlus source. | | PubMed | 1,858,966 | Open-access articles from the PubMed repository crawled in 2017. | ## Evaluation and results The model has been evaluated on the Named Entity Recognition (NER) using the following datasets: - [PharmaCoNER](https://zenodo.org/record/4270158): is a track on chemical and drug mention recognition from Spanish medical texts (for more info see: https://temu.bsc.es/pharmaconer/). - [CANTEMIST](https://zenodo.org/record/3978041#.YTt5qH2xXbQ): is a shared task specifically focusing on named entity recognition of tumor morphology, in Spanish (for more info see: https://zenodo.org/record/3978041#.YTt5qH2xXbQ). - ICTUSnet: consists of 1,006 hospital discharge reports of patients admitted for stroke from 18 different Spanish hospitals. It contains more than 79,000 annotations for 51 different kinds of variables. The evaluation results are compared against the [mBERT](https://huggingface.co/bert-base-multilingual-cased) and [BETO](https://huggingface.co/dccuchile/bert-base-spanish-wwm-cased) models: | F1 - Precision - Recall | roberta-base-biomedical-clinical-es | mBERT | BETO | |---------------------------|----------------------------|-------------------------------|-------------------------| | PharmaCoNER | **90.04** - **88.92** - **91.18** | 87.46 - 86.50 - 88.46 | 88.18 - 87.12 - 89.28 | | CANTEMIST | **83.34** - **81.48** - **85.30** | 82.61 - 81.12 - 84.15 | 82.42 - 80.91 - 84.00 | | ICTUSnet | **88.08** - **84.92** - **91.50** | 86.75 - 83.53 - 90.23 | 85.95 - 83.10 - 89.02 | ## Intended uses & limitations The model is ready-to-use only for masked language modelling to perform the Fill Mask task (try the inference API or read the next section) However, the is intended to be fine-tuned on downstream tasks such as Named Entity Recognition or Text Classification. ## Cite If you use our models, please cite our latest preprint: ```bibtex @misc{carrino2021biomedical, title={Biomedical and Clinical Language Models for Spanish: On the Benefits of Domain-Specific Pretraining in a Mid-Resource Scenario}, author={Casimiro Pio Carrino and Jordi Armengol-Estapé and Asier Gutiérrez-Fandiño and Joan Llop-Palao and Marc Pàmies and Aitor Gonzalez-Agirre and Marta Villegas}, year={2021}, eprint={2109.03570}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` If you use our Medical Crawler corpus, please cite the preprint: ```bibtex @misc{carrino2021spanish, title={Spanish Biomedical Crawled Corpus: A Large, Diverse Dataset for Spanish Biomedical Language Models}, author={Casimiro Pio Carrino and Jordi Armengol-Estapé and Ona de Gibert Bonet and Asier Gutiérrez-Fandiño and Aitor Gonzalez-Agirre and Martin Krallinger and Marta Villegas}, year={2021}, eprint={2109.07765}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` --- --- ## How to use ```python from transformers import AutoTokenizer, AutoModelForMaskedLM tokenizer = AutoTokenizer.from_pretrained("BSC-TeMU/roberta-base-biomedical-es") model = AutoModelForMaskedLM.from_pretrained("BSC-TeMU/roberta-base-biomedical-es") from transformers import pipeline unmasker = pipeline('fill-mask', model="BSC-TeMU/roberta-base-biomedical-es") unmasker("El único antecedente personal a reseñar era la <mask> arterial.") ``` ``` # Output [ { "sequence": " El único antecedente personal a reseñar era la hipertensión arterial.", "score": 0.9855039715766907, "token": 3529, "token_str": " hipertensión" }, { "sequence": " El único antecedente personal a reseñar era la diabetes arterial.", "score": 0.0039140828885138035, "token": 1945, "token_str": " diabetes" }, { "sequence": " El único antecedente personal a reseñar era la hipotensión arterial.", "score": 0.002484665485098958, "token": 11483, "token_str": " hipotensión" }, { "sequence": " El único antecedente personal a reseñar era la Hipertensión arterial.", "score": 0.0023484621196985245, "token": 12238, "token_str": " Hipertensión" }, { "sequence": " El único antecedente personal a reseñar era la presión arterial.", "score": 0.0008009297889657319, "token": 2267, "token_str": " presión" } ] ```
{"language": ["es"], "license": "apache-2.0", "tags": ["biomedical", "clinical", "spanish"], "metrics": ["ppl"], "widget": [{"text": "El \u00fanico antecedente personal a rese\u00f1ar era la <mask> arterial."}, {"text": "Las radiolog\u00edas \u00f3seas de cuerpo entero no detectan alteraciones <mask>, ni alteraciones vertebrales."}, {"text": "En el <mask> toraco-abd\u00f3mino-p\u00e9lvico no se encontraron hallazgos patol\u00f3gicos de inter\u00e9s."}]}
fill-mask
BSC-LT/roberta-base-biomedical-clinical-es
[ "transformers", "pytorch", "roberta", "fill-mask", "biomedical", "clinical", "spanish", "es", "arxiv:2109.03570", "arxiv:2109.07765", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "2109.03570", "2109.07765" ]
[ "es" ]
TAGS #transformers #pytorch #roberta #fill-mask #biomedical #clinical #spanish #es #arxiv-2109.03570 #arxiv-2109.07765 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL Biomedical-clinical language model for Spanish ============================================== Biomedical pretrained language model for Spanish. For more details about the corpus, the pretraining and the evaluation, check the official repository and read our preprint "*Carrino, C. P., Armengol-Estapé, J., Gutiérrez-Fandiño, A., Llop-Palao, J., Pàmies, M., Gonzalez-Agirre, A., & Villegas, M. (2021). Biomedical and Clinical Language Models for Spanish: On the Benefits of Domain-Specific Pretraining in a Mid-Resource Scenario.*". Tokenization and model pretraining ---------------------------------- This model is a RoBERTa-based model trained on a biomedical-clinical corpus in Spanish collected from several sources (see next section). The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original RoBERTA model with a vocabulary size of 52,000 tokens. The pretraining consists of a masked language model training at the subword level following the approach employed for the RoBERTa base model with the same hyperparameters as in the original work. The training lasted a total of 48 hours with 16 NVIDIA V100 GPUs of 16GB DDRAM, using Adam optimizer with a peak learning rate of 0.0005 and an effective batch size of 2,048 sentences. Training corpora and preprocessing ---------------------------------- The training corpus is composed of several biomedical corpora in Spanish, collected from publicly available corpora and crawlers, and a real-world clinical corpus collected from more than 278K clinical documents and notes. To obtain a high-quality training corpus while retaining the idiosyncrasies of the clinical language, a cleaning pipeline has been applied only to the biomedical corpora, keeping the clinical corpus uncleaned. Essentially, the cleaning operations used are: * data parsing in different formats + sentence splitting + language detection + filtering of ill-formed sentences + deduplication of repetitive contents + keep the original document boundaries Then, the biomedical corpora are concatenated and further global deduplication among the biomedical corpora have been applied. Eventually, the clinical corpus is concatenated to the cleaned biomedical corpus resulting in a medium-size biomedical-clinical corpus for Spanish composed of more than 1B tokens. The table below shows some basic statistics of the individual cleaned corpora: Name: Medical crawler, No. tokens: 745,705,946, Description: Crawler of more than 3,000 URLs belonging to Spanish biomedical and health domains. Name: Clinical cases misc., No. tokens: 102,855,267, Description: A miscellany of medical content, essentially clinical cases. Note that a clinical case report is a scientific publication where medical practitioners share patient cases and it is different from a clinical note or document. Name: Clinical notes/documents, No. tokens: 91,250,080, Description: Collection of more than 278K clinical documents, including discharge reports, clinical course notes and X-ray reports, for a total of 91M tokens. Name: Scielo, No. tokens: 60,007,289, Description: Publications written in Spanish crawled from the Spanish SciELO server in 2017. Name: BARR2\_background, No. tokens: 24,516,442, Description: Biomedical Abbreviation Recognition and Resolution (BARR2) containing Spanish clinical case study sections from a variety of clinical disciplines. Name: Wikipedia\_life\_sciences, No. tokens: 13,890,501, Description: Wikipedia articles crawled 04/01/2021 with the Wikipedia API python library starting from the "Ciencias\_de\_la\_vida" category up to a maximum of 5 subcategories. Multiple links to the same articles are then discarded to avoid repeating content. Name: Patents, No. tokens: 13,463,387, Description: Google Patent in Medical Domain for Spain (Spanish). The accepted codes (Medical Domain) for Json files of patents are: "A61B", "A61C","A61F", "A61H", "A61K", "A61L","A61M", "A61B", "A61P". Name: EMEA, No. tokens: 5,377,448, Description: Spanish-side documents extracted from parallel corpora made out of PDF documents from the European Medicines Agency. Name: mespen\_Medline, No. tokens: 4,166,077, Description: Spanish-side articles extracted from a collection of Spanish-English parallel corpus consisting of biomedical scientific literature. The collection of parallel resources are aggregated from the MedlinePlus source. Name: PubMed, No. tokens: 1,858,966, Description: Open-access articles from the PubMed repository crawled in 2017. Evaluation and results ---------------------- The model has been evaluated on the Named Entity Recognition (NER) using the following datasets: * PharmaCoNER: is a track on chemical and drug mention recognition from Spanish medical texts (for more info see: URL * CANTEMIST: is a shared task specifically focusing on named entity recognition of tumor morphology, in Spanish (for more info see: URL * ICTUSnet: consists of 1,006 hospital discharge reports of patients admitted for stroke from 18 different Spanish hospitals. It contains more than 79,000 annotations for 51 different kinds of variables. The evaluation results are compared against the mBERT and BETO models: Intended uses & limitations --------------------------- The model is ready-to-use only for masked language modelling to perform the Fill Mask task (try the inference API or read the next section) However, the is intended to be fine-tuned on downstream tasks such as Named Entity Recognition or Text Classification. Cite ---- If you use our models, please cite our latest preprint: If you use our Medical Crawler corpus, please cite the preprint: --- --- How to use ----------
[]
[ "TAGS\n#transformers #pytorch #roberta #fill-mask #biomedical #clinical #spanish #es #arxiv-2109.03570 #arxiv-2109.07765 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 74 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #fill-mask #biomedical #clinical #spanish #es #arxiv-2109.03570 #arxiv-2109.07765 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.07219074666500092, 0.12135610729455948, -0.007559832651168108, -0.00806131586432457, 0.06044657528400421, 0.009192046709358692, 0.12972630560398102, 0.13835851848125458, 0.03807321563363075, 0.02604307048022747, 0.21506744623184204, 0.22178199887275696, 0.02164370007812977, 0.09560072422027588, -0.019648414105176926, -0.1733534187078476, 0.050066400319337845, 0.06402367353439331, -0.04904568940401077, 0.07121523469686508, 0.07075153291225433, -0.021838421002030373, 0.04919517785310745, 0.03519998490810394, 0.004120407160371542, 0.02176268957555294, 0.11942770332098007, -0.09854673594236374, 0.1540188193321228, 0.013596086762845516, 0.12079479545354843, 0.0521310530602932, 0.020848989486694336, -0.12883637845516205, 0.013833737000823021, -0.03053053840994835, -0.06059505417943001, 0.0990818589925766, -0.021177461370825768, -0.09499900788068771, 0.07756854593753815, -0.02077065035700798, 0.012912223115563393, 0.040134575217962265, -0.128095343708992, -0.27643677592277527, -0.077431321144104, 0.06878817081451416, -0.019962361082434654, 0.08528438210487366, 0.05074484646320343, 0.17783008515834808, -0.04107602685689926, 0.034987617284059525, 0.18930357694625854, -0.31008273363113403, -0.0025669336318969727, -0.042345546185970306, 0.13677819073200226, -0.042615752667188644, 0.01884620264172554, 0.06991803646087646, 0.04723018407821655, -0.007435828447341919, 0.03552329167723656, -0.0898817777633667, -0.07918909192085266, 0.0050065042451024055, -0.04611772671341896, -0.0446476936340332, 0.16891898214817047, -0.04038672521710396, 0.009124450385570526, 0.05946546792984009, -0.06823055446147919, 0.009910318069159985, -0.009695936925709248, -0.024744611233472824, 0.047726109623909, 0.033851031213998795, 0.06849834322929382, 0.02338295616209507, -0.10053006559610367, 0.013949892483651638, -0.21835114061832428, 0.12120916694402695, -0.006799901835620403, 0.08503216505050659, -0.0646541565656662, 0.0026435672771185637, 0.029088888317346573, -0.09958691149950027, 0.03339069336652756, -0.008608698844909668, 0.07764791697263718, 0.026540108025074005, -0.027391834184527397, 0.04143121838569641, 0.15748068690299988, 0.2036011815071106, 0.01866983436048031, -0.023633655160665512, 0.02314864657819271, 0.08078940957784653, -0.05945243686437607, -0.01914507895708084, -0.019563551992177963, -0.012840792536735535, 0.038012463599443436, -0.03876831755042076, 0.08779734373092651, -0.0029209202621132135, -0.13528504967689514, -0.05063688009977341, -0.09081795811653137, 0.10585237294435501, 0.0044694934040308, -0.03649334982037544, -0.07457232475280762, 0.03043246828019619, 0.11738218367099762, -0.016414279118180275, -0.026243170723319054, -0.008856303989887238, 0.05478053167462349, 0.06594344228506088, 0.04842342808842659, 0.02216462604701519, -0.004008382558822632, 0.046630509197711945, -0.08179096132516861, -0.028149396181106567, -0.04568355530500412, -0.008983565494418144, 0.06639508903026581, -0.06662383675575256, 0.08025291562080383, -0.16529029607772827, -0.09189081192016602, 0.00813390128314495, 0.0741373747587204, -0.05223627761006355, -0.02588576264679432, 0.04623664170503616, 0.05009757727384567, 0.03833907097578049, -0.07007906585931778, -0.0006028780480846763, -0.08026458323001862, 0.09473270922899246, -0.04718426242470741, 0.08286867290735245, -0.16369450092315674, 0.000757303205318749, -0.09370262175798416, 0.021105622872710228, -0.05908619984984398, -0.12873826920986176, -0.08347165584564209, 0.09634558111429214, -0.0935201644897461, -0.03568768501281738, -0.06328710168600082, 0.006544624920934439, 0.05170046538114548, 0.09464269131422043, -0.07064014673233032, -0.07366983592510223, 0.21033744513988495, -0.038906946778297424, -0.20028607547283173, 0.05320167541503906, 0.008708903566002846, 0.07930197566747665, 0.03542386740446091, 0.1601935178041458, -0.03100094199180603, -0.20509324967861176, 0.033800508826971054, 0.005761042702943087, -0.11399969458580017, -0.19893105328083038, 0.10983727872371674, -0.10935676097869873, -0.08931127935647964, 0.028017355129122734, -0.028390496969223022, 0.0985753983259201, -0.03671697899699211, -0.06566613167524338, -0.009754653088748455, -0.08113919198513031, 0.003531886963173747, -0.010346014983952045, 0.06669899076223373, -0.0868241935968399, 0.019864937290549278, -0.05920770764350891, 0.05119902268052101, 0.09462811797857285, 0.01045083999633789, -0.10261134803295135, 0.02517898939549923, 0.004863203503191471, -0.013717010617256165, -0.09029390662908554, 0.01350557804107666, -0.018377482891082764, -0.03508751466870308, -0.022343214601278305, -0.0034969444386661053, 0.034608256071805954, -0.05891016125679016, -0.01989167556166649, 0.02267736941576004, 0.09696504473686218, 0.040597617626190186, 0.012548895552754402, -0.16245219111442566, 0.11018547415733337, -0.04903918504714966, 0.018457867205142975, 0.021850109100341797, -0.0105178477242589, -0.08275428414344788, 0.07853491604328156, -0.03085099346935749, 0.06806443631649017, -0.03462084010243416, 0.02761148475110531, -0.013697407208383083, 0.017431050539016724, 0.1326131671667099, -0.010577213019132614, -0.05298232287168503, 0.17296521365642548, -0.011506960727274418, 0.3229115903377533, 0.11843790113925934, -0.14696945250034332, 0.001298154704272747, -0.021961750462651253, -0.009571017697453499, 0.0018412413774058223, -0.002307324204593897, -0.007545212749391794, 0.04718904569745064, -0.021179279312491417, 0.12127811461687088, -0.041371285915374756, 0.03941953182220459, 0.028177393600344658, -0.1041087731719017, -0.010813532397150993, 0.048426173627376556, 0.16369354724884033, -0.16086925566196442, 0.09763065725564957, 0.2911304235458374, -0.006270596757531166, 0.12717153131961823, 0.006832837127149105, -0.009842793457210064, -0.07272770255804062, -0.06687276065349579, 0.036192845553159714, 0.12459146976470947, -0.12057565152645111, 0.02624322846531868, 0.07980895042419434, -0.0485534630715847, 0.025636259466409683, -0.08839569985866547, -0.0779879167675972, -0.0285735372453928, 0.0062796855345368385, -0.07103706896305084, 0.11269426345825195, -0.0513782873749733, 0.1111714318394661, 0.024042977020144463, -0.12084978073835373, 0.08922728896141052, 0.033847153186798096, -0.0735090970993042, 0.15619830787181854, -0.12881702184677124, -0.26406559348106384, -0.13902394473552704, -0.1190020740032196, 0.0743686705827713, 0.046016376465559006, 0.06278864294290543, -0.05083826556801796, -0.04951424524188042, 0.08739418536424637, -0.08920866996049881, 0.044558245688676834, -0.022268855944275856, -0.012074999511241913, 0.06064635142683983, 0.061984654515981674, -0.07410145550966263, -0.051227226853370667, -0.020128944888710976, 0.009170317091047764, 0.07755668461322784, -0.0898074358701706, 0.10987428575754166, 0.09257300943136215, 0.04639778286218643, -0.02928740531206131, -0.017165351659059525, 0.11266214400529861, -0.05452393367886543, -0.013323956169188023, 0.2233496755361557, 0.07071338593959808, 0.05176198109984398, 0.17201587557792664, 0.07517576217651367, -0.06803023815155029, 0.0003993251302745193, -0.05730758607387543, -0.05517656356096268, -0.2966634929180145, -0.08872907608747482, -0.06786215305328369, -0.05844283103942871, 0.02813570946455002, 0.04629794880747795, 0.1282864660024643, 0.11586876213550568, 0.0756436213850975, -0.020368337631225586, -0.13200654089450836, 0.04430483654141426, 0.20029032230377197, -0.0294375978410244, 0.1297931671142578, -0.03713122382760048, -0.09671827405691147, 0.08369498699903488, 0.0978943407535553, 0.12078790366649628, 0.14092563092708588, 0.07835233956575394, 0.1086936742067337, 0.15835939347743988, 0.061301734298467636, 0.08157076686620712, 0.09349726140499115, -0.06306158006191254, -0.049253955483436584, -0.030441084876656532, -0.08746387809515, -0.0014312247512862086, -0.029253996908664703, -0.06454560160636902, 0.0053115240298211575, -0.11389917880296707, 0.05167403072118759, 0.09505477547645569, 0.01909146085381508, -0.1446274369955063, 0.04889937862753868, 0.03831106051802635, 0.002213344443589449, -0.07892525941133499, 0.03772546723484993, -0.05859871581196785, -0.10083279758691788, 0.09418101608753204, -0.018375473096966743, 0.08298986405134201, 0.07837298512458801, 0.08816137909889221, -0.07538129389286041, -0.13669830560684204, 0.01966392993927002, 0.10036905854940414, -0.2650754749774933, 0.2913561761379242, -0.0030586873181164265, -0.019160529598593712, -0.0332469679415226, -0.014682291075587273, 0.041834041476249695, 0.18432505428791046, 0.1688011884689331, 0.05057366192340851, -0.08756841719150543, -0.03267022967338562, -0.03167027235031128, 0.04860427603125572, -0.0027616131119430065, 0.025343874469399452, -0.04942850023508072, -0.02998393028974533, -0.009863950312137604, -0.034433070570230484, 0.046645473688840866, -0.10246879607439041, -0.09398917853832245, 0.10332433879375458, 0.02569521963596344, 0.04397529363632202, -0.045957062393426895, -0.027550939470529556, -0.08591310679912567, 0.12096191942691803, -0.09046533703804016, -0.033661868423223495, -0.09489674121141434, -0.07580682635307312, 0.08356298506259918, -0.0841107964515686, 0.06304145604372025, -0.03322862833738327, -0.09164736419916153, -0.05790870636701584, -0.11622871458530426, 0.14465616643428802, -0.12745071947574615, -0.037919145077466965, -0.1343744844198227, 0.094391368329525, -0.09371167421340942, 0.05453583598136902, 0.017722534015774727, 0.018603501841425896, -0.12138841301202774, -0.044279757887125015, 0.11257844418287277, -0.07379854470491409, 0.07122493535280228, -0.02029043808579445, -0.13743330538272858, -0.14052149653434753, -0.003458653809502721, -0.08775942772626877, 0.1452610343694687, 0.236720010638237, -0.06690279394388199, 0.14238573610782623, 0.2390909492969513, -0.07775291055440903, -0.25137221813201904, -0.1412598341703415, -0.06758482754230499, -0.03775697574019432, 0.021550830453634262, -0.1820399910211563, 0.08427274227142334, 0.14463834464550018, -0.08774582296609879, 0.0681859627366066, -0.13841815292835236, -0.08488806337118149, 0.15492649376392365, 0.0034364722669124603, 0.39995086193084717, -0.1277303546667099, -0.08458180725574493, -0.053240273147821426, -0.16196280717849731, 0.09464852511882782, -0.04621828347444534, 0.05464498698711395, -0.058381397277116776, 0.0009540517930872738, -0.010563920252025127, -0.061644334346055984, 0.15737345814704895, -0.0657198578119278, 0.008571457117795944, -0.048991139978170395, -0.0706956684589386, 0.10364241153001785, 0.02590350992977619, 0.021250195801258087, -0.03961726650595665, -0.018926477059721947, -0.051324691623449326, -0.004298035055398941, -0.09491506963968277, 0.10095918923616409, 0.0046096304431557655, -0.07262733578681946, -0.05003902688622475, -0.0049860212020576, 0.005731891840696335, -0.01281408779323101, 0.198629692196846, -0.03821000084280968, 0.08230233937501907, 0.09765733778476715, 0.06378626823425293, -0.1487911343574524, -0.03338485583662987, -0.027112897485494614, -0.10530184209346771, 0.059886008501052856, -0.0888696238398552, 0.003559025702998042, 0.09833460301160812, -0.046796396374702454, 0.06174502149224281, 0.03249698877334595, -0.022587312385439873, -0.0024374164640903473, 0.1596476286649704, -0.10838880389928818, 0.05522754415869713, 0.02218250371515751, 0.12702663242816925, 0.03531967103481293, 0.034063346683979034, 0.11130926012992859, 0.04677513241767883, -0.05980833247303963, -0.00970415212213993, 0.00836267415434122, -0.10545141249895096, 0.12029294669628143, 0.09516653418540955, 0.029971305280923843, -0.07457734644412994, -0.02541981264948845, 0.016982868313789368, -0.16181892156600952, -0.010431880131363869, 0.06399677693843842, -0.11303075402975082, -0.10742117464542389, 0.015099911019206047, 0.11467298865318298, -0.24215224385261536, -0.10421766340732574, -0.1032557338476181, -0.1158095970749855, 0.052215978503227234, 0.17544212937355042, 0.03840792179107666, -0.0016946059186011553, 0.004432953894138336, -0.07082849740982056, 0.06483666598796844, 0.013046550564467907, -0.12926697731018066, 0.022870255634188652, 0.01611322909593582, -0.027498453855514526, -0.009158124215900898, 0.07602579891681671, -0.08227711170911789, 0.048482172191143036, -0.13926997780799866, 0.013227655552327633, -0.12648256123065948, -0.0014945910079404712, -0.036767031997442245, -0.074837826192379, 0.01037131529301405, -0.07531000673770905, -0.03610197827219963, -0.01947799138724804, -0.1072433665394783, -0.0012878564884886146, 0.07267072051763535, 0.05041475221514702, -0.057930972427129745, -0.0542859248816967, 0.11093797534704208, 0.008907833136618137, 0.08154269307851791, 0.06006748229265213, 0.016578836366534233, 0.06245534121990204, -0.11408527940511703, -0.053182486444711685, 0.06967940926551819, 0.09235771745443344, 0.018966130912303925, -0.07682390511035919, 0.011946936137974262, 0.10228703171014786, -0.0035228740889579058, 0.05756630003452301, 0.022520752623677254, -0.10413980484008789, 0.014472229406237602, 0.044170111417770386, -0.14470665156841278, 0.009515275247395039, -0.07691125571727753, 0.07847364246845245, -0.05531902238726616, 0.10736767202615738, -0.05776965245604515, -0.021883126348257065, -0.04957159236073494, 0.028128977864980698, -0.05473307892680168, -0.12007968872785568, -0.10634493082761765, -0.020954493433237076, -0.039759065955877304, 0.031515806913375854, 0.23202362656593323, -0.03235936909914017, -0.06613089144229889, 0.04242616891860962, 0.127081498503685, 0.04638570919632912, -0.013221572153270245, 0.17027202248573303, 0.06998001039028168, -0.007675469852983952, -0.14194409549236298, 0.055111177265644073, -0.006728761363774538, -0.058732517063617706, 0.054320819675922394, 0.0955296978354454, 0.19452963769435883, 0.004881216213107109, 0.02218254841864109, -0.009522786363959312, -0.06072758883237839, -0.13327336311340332, 0.056789763271808624, -0.023479726165533066, 0.0698600485920906, 0.024240393191576004, 0.18719960749149323, -0.027892455458641052, 0.023778298869729042, -0.039068691432476044, 0.008147838525474072, -0.14612510800361633, -0.06440580636262894, -0.03961214795708656, -0.07364997267723083, -0.013658869080245495, -0.03021792322397232, 0.0202979426831007, 0.07582735270261765, 0.0037917750887572765, -0.020976580679416656, 0.016969071701169014, -0.05110665410757065, 0.0003281092504039407, -0.006162084639072418, 0.0377434641122818, 0.03070666827261448, -0.12237592786550522, 0.023842422291636467, -0.11095208674669266, -0.002988438354805112, -0.018932446837425232, -0.009115641005337238, -0.017978494986891747, -0.037161942571401596, -0.06137915700674057, -0.031251780688762665, -0.02616867981851101, 0.017945537343621254, -0.020246373489499092, 0.15340501070022583, 0.026242950931191444, 0.053143519908189774, 0.06688214838504791, 0.14920461177825928, -0.024370502680540085, -0.09972961246967316, -0.06860494613647461, 0.08986221998929977, 0.022510351613163948, 0.08654485642910004, -0.061981599777936935, 0.0691012516617775, -0.030595697462558746, 0.32378679513931274, 0.22162505984306335, -0.07600957900285721, 0.019028570502996445, 0.01728961057960987, 0.019468098878860474, 0.06408480554819107, 0.08599637448787689, 0.11449165642261505, 0.20701077580451965, -0.08429614454507828, -0.0636318176984787, -0.08195489645004272, 0.017795143648982048, -0.18172161281108856, 0.00206818338483572, 0.00029278852161951363, -0.06405464559793472, 0.009793644770979881, 0.06092452257871628, -0.051287926733493805, 0.08458195626735687, 0.044648945331573486, -0.14701668918132782, -0.07008704543113708, -0.0368460975587368, 0.14703385531902313, 0.02860129252076149, 0.026439687237143517, -0.025438891723752022, -0.046173904091119766, 0.09295656532049179, 0.009279981255531311, -0.21267357468605042, -0.10103443264961243, 0.08236844092607498, -0.04864104464650154, 0.19037672877311707, -0.0038897532504051924, 0.029865970835089684, 0.10820073634386063, 0.04237043857574463, -0.07430098950862885, 0.11678361892700195, -0.0010415170108899474, -0.04073536768555641, 0.04327971488237381, -0.13470974564552307, 0.0373697355389595, -0.07128898799419403, 0.03683113306760788, -0.04919430986046791, 0.04565700888633728, 0.04384144768118858, -0.022304413840174675, -0.031225543469190598, 0.15935687720775604, -0.08047189563512802, 0.029074039310216904, 0.03657742589712143, 0.009148746728897095, -0.026861701160669327, -0.05049740895628929, 0.007081463001668453, 0.09579385071992874, -0.056001730263233185, -0.1137067973613739, -0.048039812594652176, 0.027302728965878487, -0.026791581884026527, -0.0015334674390032887, -0.19604988396167755, -0.0492388941347599, -0.13119100034236908, 0.001498913625255227, -0.11632461100816727, -0.023594697937369347, 0.016527026891708374, 0.02143416740000248, 0.017393026500940323, -0.05944092571735382, 0.030985044315457344, 0.015916883945465088, -0.09381103515625, -0.061517078429460526 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/roberta-base-biomedical-es # Biomedical language model for Spanish Biomedical pretrained language model for Spanish. For more details about the corpus, the pretraining and the evaluation, check the official [repository](https://github.com/PlanTL-SANIDAD/lm-biomedical-clinical-es) and read our [preprint](https://arxiv.org/abs/2109.03570) "_Carrino, C. P., Armengol-Estapé, J., Gutiérrez-Fandiño, A., Llop-Palao, J., Pàmies, M., Gonzalez-Agirre, A., & Villegas, M. (2021). Biomedical and Clinical Language Models for Spanish: On the Benefits of Domain-Specific Pretraining in a Mid-Resource Scenario._". ## Tokenization and model pretraining This model is a [RoBERTa-based](https://github.com/pytorch/fairseq/tree/master/examples/roberta) model trained on a **biomedical** corpus in Spanish collected from several sources (see next section). The training corpus has been tokenized using a byte version of [Byte-Pair Encoding (BPE)](https://github.com/openai/gpt-2) used in the original [RoBERTA](https://github.com/pytorch/fairseq/tree/master/examples/roberta) model with a vocabulary size of 52,000 tokens. The pretraining consists of a masked language model training at the subword level following the approach employed for the RoBERTa base model with the same hyperparameters as in the original work. The training lasted a total of 48 hours with 16 NVIDIA V100 GPUs of 16GB DDRAM, using Adam optimizer with a peak learning rate of 0.0005 and an effective batch size of 2,048 sentences. ## Training corpora and preprocessing The training corpus is composed of several biomedical corpora in Spanish, collected from publicly available corpora and crawlers. To obtain a high-quality training corpus, a cleaning pipeline with the following operations has been applied: - data parsing in different formats - sentence splitting - language detection - filtering of ill-formed sentences - deduplication of repetitive contents - keep the original document boundaries Finally, the corpora are concatenated and further global deduplication among the corpora have been applied. The result is a medium-size biomedical corpus for Spanish composed of about 963M tokens. The table below shows some basic statistics of the individual cleaned corpora: | Name | No. tokens | Description | |-----------------------------------------------------------------------------------------|-------------|------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | [Medical crawler](https://zenodo.org/record/4561970) | 745,705,946 | Crawler of more than 3,000 URLs belonging to Spanish biomedical and health domains. | | Clinical cases misc. | 102,855,267 | A miscellany of medical content, essentially clinical cases. Note that a clinical case report is a scientific publication where medical practitioners share patient cases and it is different from a clinical note or document. | | [Scielo](https://github.com/PlanTL-SANIDAD/SciELO-Spain-Crawler) | 60,007,289 | Publications written in Spanish crawled from the Spanish SciELO server in 2017. | | [BARR2_background](https://temu.bsc.es/BARR2/downloads/background_set.raw_text.tar.bz2) | 24,516,442 | Biomedical Abbreviation Recognition and Resolution (BARR2) containing Spanish clinical case study sections from a variety of clinical disciplines. | | Wikipedia_life_sciences | 13,890,501 | Wikipedia articles crawled 04/01/2021 with the [Wikipedia API python library](https://pypi.org/project/Wikipedia-API/) starting from the "Ciencias\_de\_la\_vida" category up to a maximum of 5 subcategories. Multiple links to the same articles are then discarded to avoid repeating content. | | Patents | 13,463,387 | Google Patent in Medical Domain for Spain (Spanish). The accepted codes (Medical Domain) for Json files of patents are: "A61B", "A61C","A61F", "A61H", "A61K", "A61L","A61M", "A61B", "A61P". | | [EMEA](http://opus.nlpl.eu/download.php?f=EMEA/v3/moses/en-es.txt.zip) | 5,377,448 | Spanish-side documents extracted from parallel corpora made out of PDF documents from the European Medicines Agency. | | [mespen_Medline](https://zenodo.org/record/3562536#.YTt1fH2xXbR) | 4,166,077 | Spanish-side articles extracted from a collection of Spanish-English parallel corpus consisting of biomedical scientific literature. The collection of parallel resources are aggregated from the MedlinePlus source. | | PubMed | 1,858,966 | Open-access articles from the PubMed repository crawled in 2017. | ## Evaluation and results The model has been evaluated on the Named Entity Recognition (NER) using the following datasets: - [PharmaCoNER](https://zenodo.org/record/4270158): is a track on chemical and drug mention recognition from Spanish medical texts (for more info see: https://temu.bsc.es/pharmaconer/). - [CANTEMIST](https://zenodo.org/record/3978041#.YTt5qH2xXbQ): is a shared task specifically focusing on named entity recognition of tumor morphology, in Spanish (for more info see: https://zenodo.org/record/3978041#.YTt5qH2xXbQ). - ICTUSnet: consists of 1,006 hospital discharge reports of patients admitted for stroke from 18 different Spanish hospitals. It contains more than 79,000 annotations for 51 different kinds of variables. The evaluation results are compared against the [mBERT](https://huggingface.co/bert-base-multilingual-cased) and [BETO](https://huggingface.co/dccuchile/bert-base-spanish-wwm-cased) models: | F1 - Precision - Recall | roberta-base-biomedical-es | mBERT | BETO | |---------------------------|----------------------------|-------------------------------|-------------------------| | PharmaCoNER | **89.48** - **87.85** - **91.18** | 87.46 - 86.50 - 88.46 | 88.18 - 87.12 - 89.28 | | CANTEMIST | **83.87** - **81.70** - **86.17** | 82.61 - 81.12 - 84.15 | 82.42 - 80.91 - 84.00 | | ICTUSnet | **88.12** - **85.56** - **90.83** | 86.75 - 83.53 - 90.23 | 85.95 - 83.10 - 89.02 | ## Intended uses & limitations The model is ready-to-use only for masked language modelling to perform the Fill Mask task (try the inference API or read the next section) However, the is intended to be fine-tuned on downstream tasks such as Named Entity Recognition or Text Classification. ## Cite If you use our models, please cite our latest preprint: ```bibtex @misc{carrino2021biomedical, title={Biomedical and Clinical Language Models for Spanish: On the Benefits of Domain-Specific Pretraining in a Mid-Resource Scenario}, author={Casimiro Pio Carrino and Jordi Armengol-Estapé and Asier Gutiérrez-Fandiño and Joan Llop-Palao and Marc Pàmies and Aitor Gonzalez-Agirre and Marta Villegas}, year={2021}, eprint={2109.03570}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` If you use our Medical Crawler corpus, please cite the preprint: ```bibtex @misc{carrino2021spanish, title={Spanish Biomedical Crawled Corpus: A Large, Diverse Dataset for Spanish Biomedical Language Models}, author={Casimiro Pio Carrino and Jordi Armengol-Estapé and Ona de Gibert Bonet and Asier Gutiérrez-Fandiño and Aitor Gonzalez-Agirre and Martin Krallinger and Marta Villegas}, year={2021}, eprint={2109.07765}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` --- ## How to use ```python from transformers import AutoTokenizer, AutoModelForMaskedLM tokenizer = AutoTokenizer.from_pretrained("BSC-TeMU/roberta-base-biomedical-es") model = AutoModelForMaskedLM.from_pretrained("BSC-TeMU/roberta-base-biomedical-es") from transformers import pipeline unmasker = pipeline('fill-mask', model="BSC-TeMU/roberta-base-biomedical-es") unmasker("El único antecedente personal a reseñar era la <mask> arterial.") ``` ``` # Output [ { "sequence": " El único antecedente personal a reseñar era la hipertensión arterial.", "score": 0.9855039715766907, "token": 3529, "token_str": " hipertensión" }, { "sequence": " El único antecedente personal a reseñar era la diabetes arterial.", "score": 0.0039140828885138035, "token": 1945, "token_str": " diabetes" }, { "sequence": " El único antecedente personal a reseñar era la hipotensión arterial.", "score": 0.002484665485098958, "token": 11483, "token_str": " hipotensión" }, { "sequence": " El único antecedente personal a reseñar era la Hipertensión arterial.", "score": 0.0023484621196985245, "token": 12238, "token_str": " Hipertensión" }, { "sequence": " El único antecedente personal a reseñar era la presión arterial.", "score": 0.0008009297889657319, "token": 2267, "token_str": " presión" } ] ```
{"language": ["es"], "license": "apache-2.0", "tags": ["biomedical", "spanish"], "metrics": ["ppl"], "widget": [{"text": "El \u00fanico antecedente personal a rese\u00f1ar era la <mask> arterial."}, {"text": "Las radiolog\u00edas \u00f3seas de cuerpo entero no detectan alteraciones <mask>, ni alteraciones vertebrales."}, {"text": "En el <mask> toraco-abd\u00f3mino-p\u00e9lvico no se encontraron hallazgos patol\u00f3gicos de inter\u00e9s."}]}
fill-mask
BSC-LT/roberta-base-biomedical-es
[ "transformers", "pytorch", "roberta", "fill-mask", "biomedical", "spanish", "es", "arxiv:2109.03570", "arxiv:2109.07765", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "2109.03570", "2109.07765" ]
[ "es" ]
TAGS #transformers #pytorch #roberta #fill-mask #biomedical #spanish #es #arxiv-2109.03570 #arxiv-2109.07765 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL Biomedical language model for Spanish ===================================== Biomedical pretrained language model for Spanish. For more details about the corpus, the pretraining and the evaluation, check the official repository and read our preprint "*Carrino, C. P., Armengol-Estapé, J., Gutiérrez-Fandiño, A., Llop-Palao, J., Pàmies, M., Gonzalez-Agirre, A., & Villegas, M. (2021). Biomedical and Clinical Language Models for Spanish: On the Benefits of Domain-Specific Pretraining in a Mid-Resource Scenario.*". Tokenization and model pretraining ---------------------------------- This model is a RoBERTa-based model trained on a biomedical corpus in Spanish collected from several sources (see next section). The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original RoBERTA model with a vocabulary size of 52,000 tokens. The pretraining consists of a masked language model training at the subword level following the approach employed for the RoBERTa base model with the same hyperparameters as in the original work. The training lasted a total of 48 hours with 16 NVIDIA V100 GPUs of 16GB DDRAM, using Adam optimizer with a peak learning rate of 0.0005 and an effective batch size of 2,048 sentences. Training corpora and preprocessing ---------------------------------- The training corpus is composed of several biomedical corpora in Spanish, collected from publicly available corpora and crawlers. To obtain a high-quality training corpus, a cleaning pipeline with the following operations has been applied: * data parsing in different formats + sentence splitting + language detection + filtering of ill-formed sentences + deduplication of repetitive contents + keep the original document boundaries Finally, the corpora are concatenated and further global deduplication among the corpora have been applied. The result is a medium-size biomedical corpus for Spanish composed of about 963M tokens. The table below shows some basic statistics of the individual cleaned corpora: Name: Medical crawler, No. tokens: 745,705,946, Description: Crawler of more than 3,000 URLs belonging to Spanish biomedical and health domains. Name: Clinical cases misc., No. tokens: 102,855,267, Description: A miscellany of medical content, essentially clinical cases. Note that a clinical case report is a scientific publication where medical practitioners share patient cases and it is different from a clinical note or document. Name: Scielo, No. tokens: 60,007,289, Description: Publications written in Spanish crawled from the Spanish SciELO server in 2017. Name: BARR2\_background, No. tokens: 24,516,442, Description: Biomedical Abbreviation Recognition and Resolution (BARR2) containing Spanish clinical case study sections from a variety of clinical disciplines. Name: Wikipedia\_life\_sciences, No. tokens: 13,890,501, Description: Wikipedia articles crawled 04/01/2021 with the Wikipedia API python library starting from the "Ciencias\_de\_la\_vida" category up to a maximum of 5 subcategories. Multiple links to the same articles are then discarded to avoid repeating content. Name: Patents, No. tokens: 13,463,387, Description: Google Patent in Medical Domain for Spain (Spanish). The accepted codes (Medical Domain) for Json files of patents are: "A61B", "A61C","A61F", "A61H", "A61K", "A61L","A61M", "A61B", "A61P". Name: EMEA, No. tokens: 5,377,448, Description: Spanish-side documents extracted from parallel corpora made out of PDF documents from the European Medicines Agency. Name: mespen\_Medline, No. tokens: 4,166,077, Description: Spanish-side articles extracted from a collection of Spanish-English parallel corpus consisting of biomedical scientific literature. The collection of parallel resources are aggregated from the MedlinePlus source. Name: PubMed, No. tokens: 1,858,966, Description: Open-access articles from the PubMed repository crawled in 2017. Evaluation and results ---------------------- The model has been evaluated on the Named Entity Recognition (NER) using the following datasets: * PharmaCoNER: is a track on chemical and drug mention recognition from Spanish medical texts (for more info see: URL * CANTEMIST: is a shared task specifically focusing on named entity recognition of tumor morphology, in Spanish (for more info see: URL * ICTUSnet: consists of 1,006 hospital discharge reports of patients admitted for stroke from 18 different Spanish hospitals. It contains more than 79,000 annotations for 51 different kinds of variables. The evaluation results are compared against the mBERT and BETO models: Intended uses & limitations --------------------------- The model is ready-to-use only for masked language modelling to perform the Fill Mask task (try the inference API or read the next section) However, the is intended to be fine-tuned on downstream tasks such as Named Entity Recognition or Text Classification. Cite ---- If you use our models, please cite our latest preprint: If you use our Medical Crawler corpus, please cite the preprint: --- How to use ----------
[]
[ "TAGS\n#transformers #pytorch #roberta #fill-mask #biomedical #spanish #es #arxiv-2109.03570 #arxiv-2109.07765 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 71 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #fill-mask #biomedical #spanish #es #arxiv-2109.03570 #arxiv-2109.07765 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.07064221054315567, 0.12206561118364334, -0.007036184426397085, 0.00028669118182733655, 0.05948682874441147, 0.006190931424498558, 0.14077028632164001, 0.13783201575279236, 0.051787905395030975, 0.022320549935102463, 0.20515064895153046, 0.23212289810180664, 0.017138157039880753, 0.10917869210243225, -0.019908448681235313, -0.17652438580989838, 0.05775866657495499, 0.05522511899471283, -0.06429670006036758, 0.07042498141527176, 0.08046161383390427, -0.025870639830827713, 0.056105900555849075, 0.033011071383953094, -0.006567642092704773, 0.02598392590880394, 0.11579965054988861, -0.10261482000350952, 0.1559543013572693, 0.01701817288994789, 0.11566342413425446, 0.04692830890417099, 0.019588500261306763, -0.1183810830116272, 0.01902989111840725, -0.025050567463040352, -0.061593107879161835, 0.10452022403478622, -0.015352582558989525, -0.08887676894664764, 0.07546808570623398, 0.0076213679276406765, 0.009669669903814793, 0.043876007199287415, -0.12452766299247742, -0.2729983329772949, -0.07583551853895187, 0.0667332112789154, -0.013249002397060394, 0.08042441308498383, 0.051517728716135025, 0.17344482243061066, -0.036626242101192474, 0.045833781361579895, 0.1995297372341156, -0.2927480936050415, 0.0009832311188802123, -0.04022865742444992, 0.1310252845287323, -0.045041657984256744, 0.014012208208441734, 0.07570500671863556, 0.056202035397291183, -0.006701321806758642, 0.034946441650390625, -0.08788615465164185, -0.06549010425806046, -0.006865301169455051, -0.04453536495566368, -0.041998911648988724, 0.16407707333564758, -0.025475189089775085, 0.006335829850286245, 0.06402267515659332, -0.07649104297161102, 0.018445324152708054, -0.010456867516040802, -0.021188687533140182, 0.04248828440904617, 0.029197417199611664, 0.0637066587805748, 0.011710630729794502, -0.1057407483458519, 0.011194737628102303, -0.2133464813232422, 0.1522929072380066, -0.003983152098953724, 0.08368738740682602, -0.06536500155925751, -0.008411788381636143, 0.019442809745669365, -0.11110111325979233, 0.03806273639202118, -0.012418908067047596, 0.09117405861616135, 0.029671424999833107, -0.03499430790543556, 0.04034879058599472, 0.15965045988559723, 0.21311388909816742, 0.0007013712311163545, -0.021184377372264862, 0.022501928731799126, 0.08834146708250046, -0.04608519747853279, -0.019176488742232323, -0.025974508374929428, -0.008788268081843853, 0.06658130884170532, -0.04584887996315956, 0.08316616714000702, -0.013320290483534336, -0.14679914712905884, -0.044915638864040375, -0.0683843120932579, 0.0964595377445221, 0.019021229818463326, -0.033685628324747086, -0.08186818659305573, 0.025235580280423164, 0.12965761125087738, -0.02628157287836075, -0.019746074452996254, -0.005770476534962654, 0.04677383601665497, 0.05238897725939751, 0.05467589199542999, 0.021699316799640656, -0.002180264098569751, 0.04351036995649338, -0.09088312089443207, -0.033625077456235886, -0.042680010199546814, -0.024547029286623, 0.07056981325149536, -0.054694049060344696, 0.08668042719364166, -0.1747734397649765, -0.09293404221534729, 0.01604451797902584, 0.07642245292663574, -0.04763638973236084, -0.04500726982951164, 0.06068988889455795, 0.04614138603210449, 0.0359506718814373, -0.07120487093925476, -0.011077090166509151, -0.07987627387046814, 0.09447330981492996, -0.039264556020498276, 0.09499777108430862, -0.16702808439731598, 0.011061843484640121, -0.09695889800786972, 0.016977274790406227, -0.07978565245866776, -0.11604318022727966, -0.09033217281103134, 0.10663589835166931, -0.09136303514242172, -0.038794245570898056, -0.05152065306901932, 0.006299985107034445, 0.05372387543320656, 0.10633654147386551, -0.060685381293296814, -0.07630398124456406, 0.20541439950466156, -0.05008102208375931, -0.19777363538742065, 0.05537274852395058, 0.014762517996132374, 0.06964781880378723, 0.038837917149066925, 0.16430330276489258, -0.0328284315764904, -0.1983896940946579, 0.055821407586336136, 0.024009911343455315, -0.0949229896068573, -0.18786658346652985, 0.10117790848016739, -0.10410658270120621, -0.09194645285606384, 0.029957272112369537, -0.017050372436642647, 0.11456036567687988, -0.03666950389742851, -0.07088856399059296, -0.007894896902143955, -0.08623146265745163, 0.027581386268138885, -0.011290466412901878, 0.06513764709234238, -0.08139235526323318, 0.020199071615934372, -0.0674959123134613, 0.048609405755996704, 0.08761347830295563, 0.01669961027801037, -0.09426652640104294, 0.04058222472667694, -0.0016194311901926994, -0.008762693963944912, -0.09857602417469025, 0.0019843156915158033, -0.014191191643476486, -0.03745526447892189, -0.013558837585151196, -0.0001781408063834533, 0.03727715089917183, -0.050880178809165955, -0.01709192991256714, 0.02489723451435566, 0.08417608588933945, 0.04468438774347305, 0.007536151446402073, -0.17766642570495605, 0.10205709934234619, -0.04193035140633583, 0.03187507018446922, 0.02620166353881359, -0.008969813585281372, -0.07513672858476639, 0.07998570799827576, -0.04337615892291069, 0.06592356413602829, -0.026565518230199814, 0.023400064557790756, -0.028171341866254807, 0.014490651898086071, 0.1270681917667389, -0.01081544253975153, -0.06751687079668045, 0.1852705329656601, -0.0154377780854702, 0.32588404417037964, 0.12958762049674988, -0.14254477620124817, -0.005772526375949383, -0.0025301408022642136, -0.00456204591318965, -0.005183025263249874, 0.006505145225673914, -0.002428837586194277, 0.02925744466483593, -0.02097778022289276, 0.12291073054075241, -0.05719974637031555, 0.04051968455314636, 0.03353649005293846, -0.11459235101938248, -0.008477449417114258, 0.04763536900281906, 0.1539488285779953, -0.17824506759643555, 0.11415204405784607, 0.2768441140651703, -0.014730141498148441, 0.11873187869787216, 0.010621474124491215, -0.00782821699976921, -0.06405788660049438, -0.06295229494571686, 0.028835611417889595, 0.1306295245885849, -0.11482106894254684, 0.03194625675678253, 0.07945432513952255, -0.04442645236849785, 0.022409826517105103, -0.09541016817092896, -0.07041960209608078, -0.024663008749485016, 0.0020768754184246063, -0.07131382077932358, 0.10799147188663483, -0.04762769490480423, 0.11160098016262054, 0.02210265025496483, -0.12206459045410156, 0.08884336054325104, 0.036054860800504684, -0.07321429997682571, 0.16964726150035858, -0.13863471150398254, -0.2643570899963379, -0.14476913213729858, -0.13115474581718445, 0.0542154498398304, 0.05086459591984749, 0.07196357846260071, -0.0559958890080452, -0.05698182433843613, 0.09018737822771072, -0.08932135254144669, 0.042952872812747955, -0.014747085981070995, -0.0006893507670611143, 0.048424553126096725, 0.06836345791816711, -0.08072959631681442, -0.050888143479824066, -0.012781075201928616, 0.019825614988803864, 0.0796239823102951, -0.0839480310678482, 0.1146194338798523, 0.0926927849650383, 0.03826797008514404, -0.03408658504486084, -0.01860761269927025, 0.14248690009117126, -0.05684242397546768, -0.008364434354007244, 0.23245376348495483, 0.0483146607875824, 0.0643586739897728, 0.15404203534126282, 0.06681171804666519, -0.07003014534711838, -0.0007113732863217592, -0.06129399687051773, -0.06853305548429489, -0.28926175832748413, -0.08710115402936935, -0.062448617070913315, -0.05029662698507309, 0.04274621233344078, 0.049233578145504, 0.14092141389846802, 0.12278828024864197, 0.05932147800922394, -0.017085295170545578, -0.12235946953296661, 0.04576396197080612, 0.19065886735916138, -0.03154736012220383, 0.12324967980384827, -0.03366423770785332, -0.08311992883682251, 0.09409908205270767, 0.09595799446105957, 0.109454445540905, 0.1400923728942871, 0.07492054253816605, 0.10302437096834183, 0.16924799978733063, 0.05868976563215256, 0.1039138063788414, 0.07524412870407104, -0.060324739664793015, -0.052086517214775085, -0.023691870272159576, -0.08914962410926819, 0.006637138314545155, -0.02655918151140213, -0.06563178449869156, 0.0033725302200764418, -0.10259934514760971, 0.04187402501702309, 0.10027410089969635, 0.016780724748969078, -0.15362223982810974, 0.032381635159254074, 0.038569580763578415, 0.00832025520503521, -0.06908925622701645, 0.031775299459695816, -0.05214496701955795, -0.10160619765520096, 0.10117728263139725, -0.017547400668263435, 0.0798984244465828, 0.0824168249964714, 0.08061987906694412, -0.07966534793376923, -0.11373419314622879, 0.02188926748931408, 0.10348740220069885, -0.26454153656959534, 0.28318729996681213, -0.007155574392527342, -0.02828383445739746, -0.04429856687784195, -0.014091216959059238, 0.044793009757995605, 0.17612560093402863, 0.15390227735042572, 0.054609738290309906, -0.090103879570961, -0.03593239560723305, -0.04963064566254616, 0.03932592645287514, -0.007558735087513924, 0.0330033116042614, -0.04751881957054138, -0.02876329980790615, -0.01248848531395197, -0.026220379397273064, 0.05916159972548485, -0.09583112597465515, -0.09392797201871872, 0.10176050662994385, 0.03084786981344223, 0.05090061202645302, -0.04682081192731857, -0.03026251494884491, -0.08205277472734451, 0.14250817894935608, -0.09443214535713196, -0.047447990626096725, -0.09864470362663269, -0.07620862126350403, 0.08973166346549988, -0.09236852079629898, 0.0730370506644249, -0.04705425724387169, -0.07409462332725525, -0.06406412273645401, -0.1224212497472763, 0.15200327336788177, -0.1336672306060791, -0.043727222830057144, -0.12289754301309586, 0.08047118782997131, -0.08419622480869293, 0.06120997667312622, 0.030135758221149445, 0.0032028253190219402, -0.12976373732089996, -0.05200595036149025, 0.10036532580852509, -0.07672526687383652, 0.05414986237883568, -0.023429393768310547, -0.14597691595554352, -0.12396670132875443, -0.008975766599178314, -0.07859639823436737, 0.1541396975517273, 0.24760279059410095, -0.06327033787965775, 0.14717942476272583, 0.25462138652801514, -0.07661227881908417, -0.2656015157699585, -0.15145014226436615, -0.0616266243159771, -0.0380316860973835, 0.023322882130742073, -0.18719221651554108, 0.08252078294754028, 0.1353917270898819, -0.09095948934555054, 0.08314941823482513, -0.1423027664422989, -0.09557125717401505, 0.16254974901676178, 0.006801382638514042, 0.4063475430011749, -0.1303068995475769, -0.07804200053215027, -0.06813737750053406, -0.1446656882762909, 0.08606525510549545, -0.02964978851377964, 0.06346261501312256, -0.04577477648854256, 0.023672068491578102, -0.011103450320661068, -0.058430589735507965, 0.1563633531332016, -0.06618937849998474, 0.004037136677652597, -0.0511748306453228, -0.06178288161754608, 0.11330179870128632, 0.022146470844745636, 0.023578718304634094, -0.042901474982500076, -0.013099616393446922, -0.048473674803972244, -0.003616510657593608, -0.09749248623847961, 0.11411166191101074, 0.000589964387472719, -0.07014936208724976, -0.03922256454825401, -0.007864370942115784, 0.005505752749741077, -0.012692168354988098, 0.2073477953672409, -0.025532878935337067, 0.09887969493865967, 0.10693724453449249, 0.05823875963687897, -0.153111532330513, -0.030458630993962288, -0.025462042540311813, -0.10240509361028671, 0.05924464389681816, -0.09217541664838791, 0.008371417410671711, 0.0867556780576706, -0.046189531683921814, 0.04983016476035118, 0.03322112187743187, -0.030408522114157677, 0.005504822824150324, 0.1594533771276474, -0.12595118582248688, 0.0629468634724617, 0.021555347368121147, 0.12134753912687302, 0.03377637267112732, 0.034405648708343506, 0.11922134459018707, 0.04621400311589241, -0.06688766926527023, -0.009595740586519241, 0.014044174924492836, -0.10581234842538834, 0.12777046859264374, 0.1019657552242279, 0.03316750004887581, -0.08661308884620667, -0.01309998705983162, 0.012735496275126934, -0.14641861617565155, 0.00009886463521979749, 0.06159014254808426, -0.11336055397987366, -0.10696688294410706, 0.0036310211289674044, 0.09716646373271942, -0.2511063516139984, -0.10120458155870438, -0.1070406585931778, -0.1143743023276329, 0.04425019025802612, 0.17816711962223053, 0.04916364327073097, -0.0007118507055565715, 0.010946649126708508, -0.07402865588665009, 0.05328480526804924, 0.004343725275248289, -0.12012440711259842, 0.02847113274037838, 0.0021247025579214096, -0.03049919195473194, -0.012147707864642143, 0.08685050904750824, -0.08218090981245041, 0.04702696204185486, -0.1460510939359665, 0.009563286788761616, -0.1263127326965332, -0.008584707975387573, -0.05074981227517128, -0.07511503249406815, 0.007250626105815172, -0.08787232637405396, -0.02747206948697567, -0.003443982219323516, -0.11471740901470184, -0.002926505170762539, 0.057598017156124115, 0.03972940519452095, -0.0649491474032402, -0.058463528752326965, 0.10926597565412521, 0.010006432421505451, 0.08056867867708206, 0.07610124349594116, 0.011045812629163265, 0.05743946135044098, -0.11377251893281937, -0.05543205142021179, 0.08223200589418411, 0.08162114024162292, 0.02627534233033657, -0.07566558569669724, 0.007143757306039333, 0.09464859217405319, -0.00019898652681149542, 0.046268608421087265, 0.041298821568489075, -0.10897453129291534, 0.009334049187600613, 0.03996383771300316, -0.1505051851272583, 0.013337872922420502, -0.07841191440820694, 0.0892869085073471, -0.059131503105163574, 0.1172524020075798, -0.05428941547870636, -0.010106615722179413, -0.04607301577925682, 0.03217383474111557, -0.05771723389625549, -0.12672311067581177, -0.12043188512325287, -0.023764975368976593, -0.04299163818359375, 0.029070358723402023, 0.2259691059589386, -0.027034781873226166, -0.0577242337167263, 0.04490085691213608, 0.11412473022937775, 0.04692724719643593, -0.011086679995059967, 0.1670909821987152, 0.06114598736166954, -0.006248370744287968, -0.12270358949899673, 0.06951946020126343, 0.00005136312393005937, -0.07129788398742676, 0.057305362075567245, 0.08475831896066666, 0.18693090975284576, 0.01352675724774599, 0.028633957728743553, -0.006351586431264877, -0.0676044374704361, -0.1425953358411789, 0.04788210615515709, -0.015307721681892872, 0.07209893316030502, 0.027171624824404716, 0.1769106239080429, -0.02741681970655918, 0.02263851836323738, -0.027455508708953857, 0.0033081250730901957, -0.1611691564321518, -0.05625780671834946, -0.049021732062101364, -0.07828196883201599, -0.012443202547729015, -0.03849884495139122, 0.014122054912149906, 0.06699515879154205, 0.0036124130710959435, -0.027357272803783417, 0.024152325466275215, -0.04078146442770958, -0.01993454433977604, -0.014893138781189919, 0.04167149215936661, 0.020313723012804985, -0.09855733811855316, 0.0063241878524422646, -0.10869162529706955, 0.005784439388662577, -0.013279563747346401, -0.007212590426206589, -0.025298243388533592, -0.020006975159049034, -0.06306735426187515, -0.041233062744140625, -0.039100661873817444, 0.02027784287929535, -0.01379459723830223, 0.13811904191970825, 0.01655610464513302, 0.044579535722732544, 0.06303244084119797, 0.1418333500623703, -0.026045382022857666, -0.10273558646440506, -0.06602015346288681, 0.11988741904497147, 0.016013627871870995, 0.08863197267055511, -0.05767281726002693, 0.06441488116979599, -0.032140105962753296, 0.3292587399482727, 0.21676722168922424, -0.0830744206905365, 0.020091727375984192, 0.009756877087056637, 0.01979913003742695, 0.057744864374399185, 0.08845805376768112, 0.11756706237792969, 0.19880573451519012, -0.0818234384059906, -0.05779169127345085, -0.08031495660543442, 0.016656704246997833, -0.19953185319900513, -0.002522320719435811, -0.0036677722819149494, -0.057538796216249466, -0.0003069362137466669, 0.061552200466394424, -0.057743459939956665, 0.07188887149095535, 0.04203176870942116, -0.1504625380039215, -0.07265431433916092, -0.03447193652391434, 0.14078699052333832, 0.0326346792280674, 0.03981815651059151, -0.026131954044103622, -0.04342273622751236, 0.0915215015411377, 0.011824121698737144, -0.20840036869049072, -0.09560643136501312, 0.08085636049509048, -0.054299596697092056, 0.18495683372020721, -0.008605977520346642, 0.028296029195189476, 0.11191762238740921, 0.03654159605503082, -0.0792044848203659, 0.09753995388746262, 0.00456220842897892, -0.05627543106675148, 0.03732951357960701, -0.13005423545837402, 0.02984817884862423, -0.0713169053196907, 0.03509778529405594, -0.04686388745903969, 0.035504814237356186, 0.039089322090148926, -0.019388698041439056, -0.020639609545469284, 0.14087456464767456, -0.08458439260721207, 0.039362892508506775, 0.03486543893814087, -0.006140934303402901, -0.037469297647476196, -0.05399095639586449, 0.0064560784958302975, 0.0994231328368187, -0.07267675548791885, -0.12188214808702469, -0.07329078018665314, 0.014158557169139385, -0.016315631568431854, 0.003629678161814809, -0.1940964311361313, -0.04206937924027443, -0.13753193616867065, 0.0029853160958737135, -0.12123914062976837, -0.0221507977694273, 0.010966942645609379, 0.018742285668849945, 0.022678665816783905, -0.057979512959718704, 0.028539791703224182, 0.021540166810154915, -0.10686434060335159, -0.06653360277414322 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/roberta-base-bne-capitel-ner-plus # Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset. RoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019. Original pre-trained model can be found here: https://huggingface.co/BSC-TeMU/roberta-base-bne ## Dataset The dataset used is the one from the [CAPITEL competition at IberLEF 2020](https://sites.google.com/view/capitel2020) (sub-task 1). **IMPORTANT ABOUT THIS MODEL:** We modified the dataset to make this model more robust to general Spanish input. In the Spanish language all the name entities are capitalized, as this dataset has been elaborated by experts, it is totally correct in terms of Spanish language. We randomly took some entities and we lower-cased some of them for the model to learn not only that the named entities are capitalized, but also the structure of a sentence that should contain a named entity. For instance: "My name is [placeholder]", this [placeholder] should be a named entity even though it is not written capitalized. The model trained on the original capitel dataset can be found here: https://huggingface.co/BSC-TeMU/roberta-base-bne-capitel-ner Examples: This model: - "Me llamo asier y vivo en barcelona todo el año." → "Me llamo {as:S-PER}{ier:S-PER} y vivo en {barcelona:S-LOC} todo el año." - "Hoy voy a visitar el parc güell tras salir del barcelona supercomputing center." → "Hoy voy a visitar el {par:B-LOC}{k:I-LOC} {gü:E-LOC}{ell:E-LOC} tras salir del {barcelona:B-ORG} {super:I-ORG}{com:I-ORG}{pu:I-ORG}{ting:I-ORG} {cen:E-ORG}{ter:E-ORG}." Model trained on original data: - "Me llamo asier y vivo en barcelona todo el año." → "Me llamo asier y vivo en barcelona todo el año." (nothing) - "Hoy voy a visitar el parc güell tras salir del barcelona supercomputing center." → "Hoy voy a visitar el parc güell tras salir del barcelona supercomputing center." (nothing) ## Evaluation and results F1 Score: 0.8867 For evaluation details visit our [GitHub repository](https://github.com/PlanTL-SANIDAD/lm-spanish). ## Citing Check out our paper for all the details: https://arxiv.org/abs/2107.07253 ``` @misc{gutierrezfandino2021spanish, title={Spanish Language Models}, author={Asier Gutiérrez-Fandiño and Jordi Armengol-Estapé and Marc Pàmies and Joan Llop-Palao and Joaquín Silveira-Ocampo and Casimiro Pio Carrino and Aitor Gonzalez-Agirre and Carme Armentano-Oller and Carlos Rodriguez-Penagos and Marta Villegas}, year={2021}, eprint={2107.07253}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
{"language": ["es"], "license": "apache-2.0", "tags": ["national library of spain", "spanish", "bne", "capitel", "ner"], "datasets": ["bne", "capitel"], "metrics": ["f1"], "inference": {"parameters": {"aggregation_strategy": "first"}}}
token-classification
BSC-LT/roberta-base-bne-capitel-ner-plus
[ "transformers", "pytorch", "roberta", "token-classification", "national library of spain", "spanish", "bne", "capitel", "ner", "es", "dataset:bne", "dataset:capitel", "arxiv:1907.11692", "arxiv:2107.07253", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1907.11692", "2107.07253" ]
[ "es" ]
TAGS #transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #ner #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL # Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset. RoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019. Original pre-trained model can be found here: URL ## Dataset The dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 1). IMPORTANT ABOUT THIS MODEL: We modified the dataset to make this model more robust to general Spanish input. In the Spanish language all the name entities are capitalized, as this dataset has been elaborated by experts, it is totally correct in terms of Spanish language. We randomly took some entities and we lower-cased some of them for the model to learn not only that the named entities are capitalized, but also the structure of a sentence that should contain a named entity. For instance: "My name is [placeholder]", this [placeholder] should be a named entity even though it is not written capitalized. The model trained on the original capitel dataset can be found here: URL Examples: This model: - "Me llamo asier y vivo en barcelona todo el año." → "Me llamo {as:S-PER}{ier:S-PER} y vivo en {barcelona:S-LOC} todo el año." - "Hoy voy a visitar el parc güell tras salir del barcelona supercomputing center." → "Hoy voy a visitar el {par:B-LOC}{k:I-LOC} {gü:E-LOC}{ell:E-LOC} tras salir del {barcelona:B-ORG} {super:I-ORG}{com:I-ORG}{pu:I-ORG}{ting:I-ORG} {cen:E-ORG}{ter:E-ORG}." Model trained on original data: - "Me llamo asier y vivo en barcelona todo el año." → "Me llamo asier y vivo en barcelona todo el año." (nothing) - "Hoy voy a visitar el parc güell tras salir del barcelona supercomputing center." → "Hoy voy a visitar el parc güell tras salir del barcelona supercomputing center." (nothing) ## Evaluation and results F1 Score: 0.8867 For evaluation details visit our GitHub repository. ## Citing Check out our paper for all the details: URL
[ "# Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset.\nRoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 1).\n\nIMPORTANT ABOUT THIS MODEL: We modified the dataset to make this model more robust to general Spanish input. In the Spanish language all the name entities are capitalized, as this dataset has been elaborated by experts, it is totally correct in terms of Spanish language. We randomly took some entities and we lower-cased some of them for the model to learn not only that the named entities are capitalized, but also the structure of a sentence that should contain a named entity. For instance: \"My name is [placeholder]\", this [placeholder] should be a named entity even though it is not written capitalized. The model trained on the original capitel dataset can be found here: URL\n\nExamples:\n\nThis model:\n- \"Me llamo asier y vivo en barcelona todo el año.\" → \"Me llamo {as:S-PER}{ier:S-PER} y vivo en {barcelona:S-LOC} todo el año.\"\n- \"Hoy voy a visitar el parc güell tras salir del barcelona supercomputing center.\" → \"Hoy voy a visitar el {par:B-LOC}{k:I-LOC} {gü:E-LOC}{ell:E-LOC} tras salir del {barcelona:B-ORG} {super:I-ORG}{com:I-ORG}{pu:I-ORG}{ting:I-ORG} {cen:E-ORG}{ter:E-ORG}.\"\n\nModel trained on original data:\n- \"Me llamo asier y vivo en barcelona todo el año.\" → \"Me llamo asier y vivo en barcelona todo el año.\" (nothing)\n- \"Hoy voy a visitar el parc güell tras salir del barcelona supercomputing center.\" → \"Hoy voy a visitar el parc güell tras salir del barcelona supercomputing center.\" (nothing)", "## Evaluation and results\nF1 Score: 0.8867\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ "TAGS\n#transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #ner #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset.\nRoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 1).\n\nIMPORTANT ABOUT THIS MODEL: We modified the dataset to make this model more robust to general Spanish input. In the Spanish language all the name entities are capitalized, as this dataset has been elaborated by experts, it is totally correct in terms of Spanish language. We randomly took some entities and we lower-cased some of them for the model to learn not only that the named entities are capitalized, but also the structure of a sentence that should contain a named entity. For instance: \"My name is [placeholder]\", this [placeholder] should be a named entity even though it is not written capitalized. The model trained on the original capitel dataset can be found here: URL\n\nExamples:\n\nThis model:\n- \"Me llamo asier y vivo en barcelona todo el año.\" → \"Me llamo {as:S-PER}{ier:S-PER} y vivo en {barcelona:S-LOC} todo el año.\"\n- \"Hoy voy a visitar el parc güell tras salir del barcelona supercomputing center.\" → \"Hoy voy a visitar el {par:B-LOC}{k:I-LOC} {gü:E-LOC}{ell:E-LOC} tras salir del {barcelona:B-ORG} {super:I-ORG}{com:I-ORG}{pu:I-ORG}{ting:I-ORG} {cen:E-ORG}{ter:E-ORG}.\"\n\nModel trained on original data:\n- \"Me llamo asier y vivo en barcelona todo el año.\" → \"Me llamo asier y vivo en barcelona todo el año.\" (nothing)\n- \"Hoy voy a visitar el parc güell tras salir del barcelona supercomputing center.\" → \"Hoy voy a visitar el parc güell tras salir del barcelona supercomputing center.\" (nothing)", "## Evaluation and results\nF1 Score: 0.8867\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ 95, 142, 452, 24, 13 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #ner #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset.\nRoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL" ]
[ -0.1018202155828476, 0.16918718814849854, -0.0035955526400357485, 0.03688250109553337, 0.04522613808512688, -0.041523683816194534, 0.022967472672462463, 0.05426777899265289, -0.061703871935606, 0.0005459602689370513, 0.06933991611003876, 0.12060613930225372, 0.003006361424922943, 0.011753007769584656, 0.025161931291222572, -0.2749863564968109, 0.06093151122331619, 0.009702612645924091, -0.008641541935503483, 0.07650187611579895, 0.06853267550468445, -0.07185950875282288, -0.012073338963091373, 0.0073715983889997005, -0.08513470739126205, 0.0816991999745369, -0.04243018105626106, -0.18915767967700958, 0.11010842770338058, 0.005415476858615875, 0.1489131599664688, 0.04785626381635666, 0.03717717528343201, 0.018773047253489494, -0.00008246857032645494, 0.0029727150686085224, -0.017379041761159897, 0.06225837394595146, 0.09760717302560806, -0.11678222566843033, 0.17918159067630768, 0.019399018958210945, 0.00539410300552845, 0.016862839460372925, -0.18553248047828674, -0.232735276222229, -0.05937493219971657, -0.019888965412974358, 0.0509679913520813, 0.0854632630944252, -0.020001672208309174, 0.06351052969694138, -0.07367406040430069, -0.0017849293071776628, 0.02478409744799137, -0.1660541445016861, -0.06947214901447296, -0.0077049629762768745, 0.0438922718167305, 0.05454699695110321, -0.00036233736318536103, 0.033661115914583206, 0.0645274668931961, 0.018064191564917564, 0.005107766482979059, -0.10844215750694275, -0.16591429710388184, -0.0401175320148468, -0.1324424296617508, -0.050871871411800385, 0.06663551181554794, -0.0921965017914772, -0.03575441241264343, 0.03891602158546448, -0.09031444787979126, 0.06428203731775284, -0.028577912598848343, -0.03901173919439316, 0.05622127279639244, -0.05281506106257439, 0.044823095202445984, -0.06581077724695206, -0.03980647027492523, -0.03031584434211254, -0.15067912638187408, 0.24633805453777313, 0.027099458500742912, 0.01335871685296297, -0.0031877397559583187, 0.03651229664683342, 0.07777748256921768, -0.06793644279241562, 0.007143654394894838, -0.010623464360833168, 0.020206864923238754, 0.047211118042469025, -0.08605578541755676, -0.1337941735982895, -0.007382635027170181, 0.0882074236869812, -0.0964241549372673, -0.0372483991086483, -0.025658486410975456, 0.05737279728055, 0.05265696346759796, 0.147991344332695, -0.05788995325565338, -0.09247703105211258, 0.0520053505897522, -0.040235355496406555, 0.04118495434522629, 0.006063215900212526, -0.1668025106191635, -0.1220439225435257, 0.006899590604007244, 0.06303942203521729, 0.09497427940368652, 0.0017886386485770345, -0.0028310983907431364, -0.08950664848089218, 0.10735233873128891, -0.07851803302764893, -0.06442166119813919, -0.040233854204416275, -0.10132266581058502, -0.012495044618844986, 0.010779568925499916, -0.03989674896001816, -0.07910819351673126, -0.10325036197900772, -0.0555243082344532, -0.11135830730199814, -0.10658050328493118, -0.12710750102996826, 0.04350100830197334, -0.1428680568933487, 0.0020811567083001137, -0.15405841171741486, -0.13452517986297607, -0.08394868671894073, 0.0158693827688694, -0.06509168446063995, 0.04382502660155296, -0.11152160912752151, 0.08242474496364594, -0.01548156887292862, -0.060946229845285416, 0.03310130164027214, -0.019658628851175308, 0.06601405888795853, 0.043273668736219406, 0.09533639252185822, -0.16798309981822968, 0.04189697280526161, -0.1734161078929901, -0.004231628030538559, -0.2145327478647232, 0.11185164749622345, -0.02980295941233635, 0.00445070443674922, -0.12026228755712509, -0.07379382848739624, -0.0748860090970993, 0.1040298119187355, 0.06498107314109802, 0.08381292968988419, -0.05987400561571121, -0.08003135025501251, 0.19479161500930786, -0.05968651920557022, -0.00717973941937089, 0.057624559849500656, -0.01193854957818985, 0.1333755999803543, 0.09737610071897507, 0.1524585485458374, 0.05183551087975502, 0.05707436054944992, 0.032655902206897736, -0.020571459084749222, 0.06718264520168304, -0.18039417266845703, 0.08552494645118713, -0.08292414247989655, -0.04339897632598877, 0.03635462746024132, -0.1615559309720993, 0.03979878127574921, -0.0063577196560800076, -0.017239060252904892, 0.0968804582953453, -0.01281085517257452, 0.043112386018037796, 0.047975048422813416, 0.09871252626180649, -0.008792911656200886, -0.02787797711789608, 0.10962557792663574, 0.021455192938447, 0.0063422443345189095, 0.020799236372113228, -0.048438724130392075, 0.11809363216161728, -0.11330927908420563, 0.016324171796441078, -0.1987563669681549, 0.024939924478530884, -0.004670977126806974, 0.04637623950839043, 0.06156406179070473, 0.012976747937500477, 0.006286787800490856, -0.0002653129631653428, -0.0049406117759644985, -0.00047558805090375245, -0.005401124712079763, -0.05782175436615944, -0.002847417024895549, -0.10410559177398682, -0.015554782003164291, -0.049518898129463196, 0.009234173223376274, -0.03987849876284599, -0.023746561259031296, -0.12862040102481842, 0.013434765860438347, -0.05748641490936279, 0.027602391317486763, 0.006696674041450024, 0.13596124947071075, -0.04956730082631111, -0.0059868330135941505, 0.09054698050022125, 0.016795001924037933, -0.014369434677064419, 0.1264917254447937, -0.07134292274713516, -0.02015453390777111, 0.12234140932559967, -0.17432858049869537, 0.012697055004537106, 0.0392620787024498, -0.02552434243261814, 0.03466566652059555, 0.07702361792325974, -0.010350748896598816, 0.26899075508117676, -0.003288217820227146, 0.13114085793495178, -0.18698062002658844, -0.008040888234972954, -0.005097643006592989, -0.046738263219594955, -0.04856501519680023, 0.14446911215782166, 0.03940243646502495, -0.14799542725086212, 0.0946849137544632, 0.04015723615884781, -0.06205376237630844, 0.20556510984897614, 0.036549847573041916, -0.029338475316762924, 0.03495696932077408, -0.04902295023202896, 0.003791038878262043, 0.09367576986551285, -0.12900415062904358, -0.07004906237125397, 0.0025467160157859325, -0.01889726147055626, 0.09239568561315536, -0.08430436998605728, 0.009445494040846825, -0.025638312101364136, -0.04347172752022743, -0.028259487822651863, 0.020686106756329536, -0.046904031187295914, 0.09152814000844955, 0.09678028523921967, -0.11675536632537842, -0.039088234305381775, -0.014908188953995705, -0.050991520285606384, 0.1735481470823288, -0.12275312095880508, -0.25182512402534485, -0.1464337259531021, -0.043296512216329575, -0.042227067053318024, 0.1196238249540329, -0.001700147520750761, -0.08128988742828369, 0.032763272523880005, 0.05697490647435188, 0.14180220663547516, -0.09587611258029938, -0.08469764143228531, -0.0688723474740982, 0.016729097813367844, -0.08392040431499481, -0.1420421153306961, 0.022897621616721153, -0.07411611080169678, -0.07782579958438873, -0.01177274901419878, -0.11672953516244888, 0.1631830632686615, 0.11918573081493378, -0.002028601011261344, -0.045865483582019806, -0.045271068811416626, 0.11049220710992813, -0.08082831650972366, -0.03243960067629814, 0.16690316796302795, 0.06427855789661407, -0.03161627799272537, 0.05171654000878334, -0.004101607017219067, -0.04155203700065613, -0.05862699821591377, 0.017178745940327644, -0.08471948653459549, -0.21601250767707825, -0.15591591596603394, -0.032093945890665054, -0.07521989941596985, 0.11208756268024445, 0.002786024007946253, 0.05890268087387085, 0.08425071835517883, -0.03372962772846222, 0.017443766817450523, -0.022095218300819397, 0.05698743090033531, 0.1709766536951065, -0.016363734379410744, 0.08871462196111679, -0.013696492649614811, -0.07338324189186096, 0.07511792331933975, 0.05619063973426819, 0.1177113875746727, -0.01250902097672224, 0.12733542919158936, 0.06987785547971725, 0.08505705744028091, 0.016183393076062202, 0.07527687400579453, -0.017290037125349045, 0.0817350521683693, -0.07732533663511276, -0.052008479833602905, -0.11988812685012817, -0.06066027656197548, -0.041893694549798965, -0.0006118294550105929, -0.09727746248245239, -0.16787156462669373, 0.010752441361546516, 0.14157021045684814, -0.007120385300368071, -0.2540181577205658, -0.050137948244810104, 0.025929754599928856, 0.014742347411811352, -0.09857569634914398, 0.03196777403354645, -0.05290816351771355, -0.07469050586223602, 0.05540269985795021, 0.012844134122133255, 0.10666157305240631, -0.09969881922006607, 0.038372915238142014, -0.09772033244371414, 0.039108775556087494, 0.03577690199017525, 0.09009687602519989, -0.10962390154600143, 0.3031950294971466, 0.02244146354496479, 0.03857472166419029, -0.018675396218895912, -0.01656077802181244, -0.03647366538643837, 0.008447394706308842, 0.12040365487337112, 0.019977319985628128, 0.019442982971668243, -0.02722018212080002, -0.04654031619429588, 0.036658741533756256, -0.03444940596818924, -0.050351954996585846, 0.09040030837059021, 0.02405107207596302, -0.021587125957012177, -0.01992405764758587, -0.10339918732643127, -0.0381743386387825, -0.08684306591749191, -0.0002379369834670797, -0.04020582139492035, 0.009853563271462917, -0.014699245803058147, -0.05967367812991142, -0.011263067834079266, 0.10054214298725128, -0.04099244624376297, -0.06049627438187599, -0.06833712756633759, 0.060648057609796524, 0.14295758306980133, -0.06633001565933228, 0.0458381250500679, -0.05215265229344368, 0.0016825676430016756, -0.03207552060484886, -0.09703810513019562, 0.09204447269439697, -0.1406174600124359, 0.011619796976447105, -0.03492337465286255, 0.02144438773393631, 0.07367902994155884, 0.015638692304491997, 0.06659390032291412, 0.016098270192742348, -0.01591111719608307, -0.05237461254000664, -0.0348799042403698, -0.04499103128910065, 0.060584601014852524, 0.03018919751048088, -0.15897250175476074, -0.1046881452202797, -0.022424817085266113, 0.060647305101156235, 0.1037546694278717, 0.021460147574543953, -0.01063043624162674, 0.08488254994153976, 0.3597245216369629, -0.1334788054227829, -0.22288045287132263, -0.03000352531671524, 0.07038427144289017, 0.05864255875349045, -0.16623912751674652, -0.2882053554058075, 0.013805894181132317, 0.13663966953754425, -0.02849111147224903, -0.03330853208899498, -0.3138817548751831, -0.04122116044163704, 0.17556791007518768, 0.05924099311232567, 0.29987162351608276, -0.0473729632794857, -0.066328264772892, -0.05846119672060013, 0.09337712079286575, 0.07799164205789566, -0.048668887466192245, 0.03848061338067055, -0.01503361202776432, -0.03809370845556259, 0.02624398097395897, 0.009305856190621853, 0.10666509717702866, -0.051999032497406006, -0.004783253185451031, -0.005600155331194401, -0.002230739686638117, 0.14988648891448975, 0.008408709429204464, 0.015118944458663464, 0.15741917490959167, 0.007145268376916647, -0.13198259472846985, -0.061443429440259933, -0.10639641433954239, 0.06183791533112526, -0.022167304530739784, -0.030485916882753372, 0.038198597729206085, 0.0535101518034935, 0.013406205922365189, 0.014514341950416565, 0.07723823934793472, -0.08214927464723587, 0.09702616184949875, 0.09964518994092941, 0.16012947261333466, 0.039359159767627716, 0.05939390882849693, 0.032500386238098145, -0.011169771663844585, 0.09353575855493546, 0.04927954450249672, -0.03768329322338104, 0.09920932352542877, -0.011139016598463058, -0.04198795557022095, 0.0341976173222065, -0.1248774304986, 0.004491180647164583, 0.13629218935966492, -0.018320580944418907, 0.022988351061940193, -0.027723975479602814, -0.07864826172590256, 0.026525497436523438, -0.0016866473015397787, 0.14954188466072083, 0.011223939247429371, -0.09441706538200378, -0.010491210035979748, -0.05047253519296646, 0.0006501656025648117, 0.09445185214281082, 0.040414806455373764, -0.014216274954378605, -0.0906905010342598, 0.09266089648008347, 0.11899096518754959, -0.03244401514530182, 0.006304537877440453, 0.0991763100028038, -0.04526817798614502, -0.07290829718112946, -0.007155518047511578, 0.11930068582296371, -0.31168046593666077, -0.0898727998137474, -0.1152115911245346, -0.07782614231109619, 0.01732882671058178, 0.25210484862327576, 0.037853509187698364, -0.029808998107910156, -0.027772748842835426, 0.03790351375937462, -0.011131487786769867, -0.018573058769106865, -0.008067460730671883, 0.0006187228718772531, 0.03793427348136902, 0.09577513486146927, 0.04769095405936241, 0.0009822694119066, -0.05136805772781372, -0.03193120285868645, -0.12831845879554749, 0.05206310376524925, -0.1530715674161911, 0.0010994401527568698, -0.054796021431684494, -0.024074183776974678, -0.07540201395750046, -0.004721507430076599, -0.034156110137701035, -0.05227378010749817, -0.07445650547742844, 0.02429213374853134, -0.006167108193039894, 0.009764221496880054, -0.05493329465389252, 0.002039415994659066, -0.03616148233413696, 0.0016339324647560716, 0.0006450192886404693, 0.03926673159003258, 0.03178377449512482, 0.121106818318367, -0.11345156282186508, 0.05094052851200104, 0.043795906007289886, 0.0446581169962883, 0.055602315813302994, 0.010298192501068115, 0.030061157420277596, 0.0800044983625412, -0.005147529300302267, 0.037207238376140594, -0.03514654561877251, -0.10696829855442047, 0.017987769097089767, 0.08306069672107697, -0.06435184925794601, -0.014483971521258354, 0.12340997904539108, 0.08859124779701233, 0.06519302725791931, 0.035604868084192276, -0.07278762757778168, 0.009249404072761536, -0.073903888463974, -0.002571983030065894, -0.05998436361551285, -0.028589051216840744, -0.1022181287407875, -0.03783941641449928, 0.022906184196472168, 0.04670765995979309, 0.17314353585243225, 0.10595937818288803, 0.1494555026292801, -0.025379033759236336, -0.011427750810980797, -0.024892371147871017, 0.008183164522051811, 0.15676645934581757, 0.11603615432977676, -0.0017706905491650105, 0.01918027177453041, 0.14502514898777008, 0.01646169275045395, 0.07459043711423874, 0.03090818226337433, 0.16997572779655457, 0.3100588917732239, 0.11079485714435577, 0.06550361961126328, -0.007437811233103275, 0.0350070521235466, 0.020881731063127518, 0.07363623380661011, 0.011441139504313469, 0.03841442987322807, 0.06395671516656876, 0.107501320540905, -0.093046173453331, 0.08392910659313202, 0.09044332802295685, 0.007575254421681166, -0.11982136964797974, -0.12658782303333282, -0.00293914508074522, -0.04525655135512352, -0.043756067752838135, -0.1128825843334198, -0.06032739579677582, -0.08096826821565628, 0.009082164615392685, -0.0641871839761734, 0.012585545890033245, -0.0069201611913740635, -0.186570942401886, 0.08170750737190247, 0.02023758366703987, 0.20374323427677155, -0.060659706592559814, 0.0029409064445644617, -0.008895132690668106, 0.15236349403858185, 0.01334844995290041, 0.03023683838546276, -0.009729568846523762, 0.061796884983778, -0.03247835114598274, -0.03291726112365723, 0.005419777240604162, 0.06575265526771545, 0.10543467849493027, 0.19115889072418213, 0.03164088726043701, -0.08022306114435196, 0.005200921557843685, 0.21696285903453827, 0.015194184146821499, -0.009687612764537334, -0.02806243672966957, 0.22629791498184204, 0.03736437112092972, 0.05764472484588623, -0.023304689675569534, -0.06118975579738617, -0.04277767986059189, 0.23250535130500793, 0.2596098482608795, 0.002168246079236269, -0.050409745424985886, 0.028287498280405998, -0.00045562293962575495, 0.09042643010616302, 0.1079900786280632, 0.02891993708908558, 0.37702736258506775, -0.03681986406445503, -0.09045344591140747, -0.0604407824575901, 0.08102025091648102, -0.05880892276763916, 0.09638995677232742, -0.009892505593597889, -0.08062881231307983, -0.018079271540045738, 0.11069708317518234, -0.13091692328453064, -0.2188018560409546, 0.03829130157828331, -0.12406763434410095, -0.1389697790145874, -0.06214294955134392, -0.026529785245656967, 0.10080746561288834, 0.1037774384021759, 0.02050209604203701, -0.03462870419025421, -0.009851311333477497, 0.02477942779660225, -0.10357704758644104, -0.20909501612186432, 0.09227052330970764, 0.10977856814861298, 0.24387532472610474, -0.052274566143751144, -0.0017330488190054893, 0.0735514909029007, 0.037398677319288254, -0.06935718655586243, 0.027365149930119514, 0.006752555258572102, 0.013014431111514568, 0.13008208572864532, -0.039234381169080734, -0.00941124465316534, 0.051365409046411514, 0.044229310005903244, -0.0789182186126709, 0.07601766288280487, 0.03004658967256546, -0.09150021523237228, -0.06334943324327469, 0.08608207106590271, -0.11310497671365738, 0.06955146044492722, 0.18544384837150574, -0.002944695297628641, -0.0030700659845024347, -0.06281744688749313, 0.0430266298353672, 0.02188858576118946, 0.07969047129154205, -0.06424514204263687, -0.14795368909835815, -0.05491640791296959, -0.008927715942263603, -0.03545869514346123, -0.2733393609523773, 0.09277835488319397, -0.006905725691467524, -0.03444932773709297, -0.024676481261849403, 0.02066662907600403, -0.07123097777366638, 0.0446138009428978, 0.010028750635683537, 0.09271464496850967, -0.019916178658604622, 0.044580552726984024, -0.11987846344709396, -0.06571843475103378 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/roberta-base-bne-capitel-ner # Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset. RoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019. Original pre-trained model can be found here: https://huggingface.co/BSC-TeMU/roberta-base-bne ## Dataset The dataset used is the one from the [CAPITEL competition at IberLEF 2020](https://sites.google.com/view/capitel2020) (sub-task 1). ## Evaluation and results F1 Score: 0.8960 For evaluation details visit our [GitHub repository](https://github.com/PlanTL-SANIDAD/lm-spanish). ## Citing Check out our paper for all the details: https://arxiv.org/abs/2107.07253 ``` @misc{gutierrezfandino2021spanish, title={Spanish Language Models}, author={Asier Gutiérrez-Fandiño and Jordi Armengol-Estapé and Marc Pàmies and Joan Llop-Palao and Joaquín Silveira-Ocampo and Casimiro Pio Carrino and Aitor Gonzalez-Agirre and Carme Armentano-Oller and Carlos Rodriguez-Penagos and Marta Villegas}, year={2021}, eprint={2107.07253}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
{"language": ["es"], "license": "apache-2.0", "tags": ["national library of spain", "spanish", "bne", "capitel", "ner"], "datasets": ["bne", "capitel"], "metrics": ["f1"]}
token-classification
BSC-LT/roberta-base-bne-capitel-ner
[ "transformers", "pytorch", "roberta", "token-classification", "national library of spain", "spanish", "bne", "capitel", "ner", "es", "dataset:bne", "dataset:capitel", "arxiv:1907.11692", "arxiv:2107.07253", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1907.11692", "2107.07253" ]
[ "es" ]
TAGS #transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #ner #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL # Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset. RoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019. Original pre-trained model can be found here: URL ## Dataset The dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 1). ## Evaluation and results F1 Score: 0.8960 For evaluation details visit our GitHub repository. ## Citing Check out our paper for all the details: URL
[ "# Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset.\nRoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 1).", "## Evaluation and results\nF1 Score: 0.8960\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ "TAGS\n#transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #ner #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset.\nRoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 1).", "## Evaluation and results\nF1 Score: 0.8960\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ 95, 142, 29, 24, 13 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #ner #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset.\nRoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 1).## Evaluation and results\nF1 Score: 0.8960\n\nFor evaluation details visit our GitHub repository.## Citing \nCheck out our paper for all the details: URL" ]
[ -0.14008329808712006, 0.23199720680713654, -0.005214326083660126, 0.04283278062939644, 0.050990864634513855, -0.02942052111029625, -0.007053106091916561, 0.09038899093866348, -0.05680089816451073, 0.0852636843919754, 0.008230399340391159, 0.11120045930147171, 0.0830392986536026, 0.06354713439941406, 0.002654426731169224, -0.17342746257781982, 0.04926187917590141, -0.020498180761933327, 0.03409066051244736, 0.07005121558904648, 0.08540522307157516, -0.05922135338187218, 0.0026731581892818213, 0.022180166095495224, -0.049007348716259, 0.09221039712429047, -0.03203858435153961, -0.17627239227294922, 0.08541477471590042, 0.02817079797387123, 0.11757524311542511, 0.04018465802073479, 0.0011916827643290162, -0.024875519797205925, -0.005477354396134615, 0.030824633315205574, -0.008246287703514099, 0.054531313478946686, 0.13743002712726593, -0.12672144174575806, 0.21435078978538513, -0.010789785534143448, 0.013078335672616959, 0.042336974292993546, -0.18163980543613434, -0.28157609701156616, -0.1368207335472107, -0.009853025898337364, 0.06772781908512115, 0.08284422010183334, -0.035527538508176804, 0.11154977232217789, -0.04763485863804817, -0.015005819499492645, 0.034918129444122314, -0.1682557314634323, -0.05569201335310936, 0.027009783312678337, 0.017078332602977753, 0.07653537392616272, -0.04249804466962814, 0.00570320151746273, 0.09646463394165039, 0.009159093722701073, -0.0410565584897995, -0.08032143861055374, -0.11889570951461792, -0.020033856853842735, -0.12324890494346619, -0.08708178251981735, 0.10266179591417313, -0.06965020298957825, -0.08317231386899948, -0.019901098683476448, -0.04889731481671333, 0.0234878808259964, 0.013301735743880272, -0.027050433680415154, 0.06949394941329956, -0.0332811065018177, 0.06789160519838333, -0.004695539828389883, -0.07232652604579926, -0.015953149646520615, -0.12249482423067093, 0.1637897491455078, 0.00517024053260684, 0.02259296551346779, 0.015813641250133514, 0.06865200400352478, 0.007499143946915865, -0.09959667921066284, -0.06344001740217209, -0.004833745304495096, -0.02548731490969658, -0.03078722208738327, -0.03527794033288956, -0.14083081483840942, 0.014699029736220837, 0.0688503086566925, -0.17002899944782257, -0.02055969089269638, -0.0361781120300293, 0.013376540504395962, 0.12242547422647476, 0.16741125285625458, -0.03177165985107422, -0.0897175520658493, -0.005718897562474012, -0.05952819064259529, 0.033112332224845886, 0.025502141565084457, -0.05069825053215027, -0.08484348654747009, 0.016545221209526062, 0.09311956912279129, 0.07636665552854538, -0.06985251605510712, -0.055013809353113174, -0.05177523195743561, 0.16735462844371796, -0.10286476463079453, -0.022221790626645088, -0.024623621255159378, -0.11180375516414642, 0.027245940640568733, -0.021727073937654495, -0.053256817162036896, -0.07844407111406326, -0.053069230169057846, -0.08014838397502899, -0.1147499829530716, -0.09045328944921494, -0.09670725464820862, 0.08181646466255188, -0.130064457654953, 0.0016109808348119259, -0.10489131510257721, -0.06127409264445305, -0.12922193109989166, -0.007910369895398617, -0.09036584943532944, 0.06179755553603172, -0.07782547920942307, 0.057880956679582596, -0.003982567694038153, -0.02298334240913391, 0.04673072323203087, -0.02723497897386551, 0.0692753717303276, 0.07926427572965622, 0.0613485649228096, -0.08379199355840683, 0.0157951470464468, -0.17341922223567963, -0.0023469016887247562, -0.23208342492580414, 0.1219279021024704, -0.08579326421022415, -0.0002955996314994991, -0.15911231935024261, -0.06873331218957901, -0.05896741896867752, 0.06242313235998154, 0.06625675410032272, 0.11791816353797913, -0.0781133845448494, -0.08261669427156448, 0.1945110708475113, -0.1139792650938034, -0.04926168546080589, 0.05666758120059967, 0.006296281237155199, 0.08037921786308289, 0.10532929003238678, 0.10992211103439331, 0.08904120326042175, -0.017972582951188087, -0.09365235269069672, -0.012984812259674072, 0.03150738775730133, -0.10826840251684189, 0.11614349484443665, -0.11921123415231705, 0.04459210857748985, 0.02538568153977394, -0.15706902742385864, -0.012439643032848835, -0.006859702989459038, -0.014662797562777996, 0.06197173148393631, 0.01883550174534321, -0.06430062651634216, 0.0599081851541996, 0.03367382287979126, -0.03332610800862312, -0.03577068820595741, 0.04307027906179428, 0.046594928950071335, 0.03058227151632309, -0.00024731067242100835, -0.05714474990963936, 0.14217312633991241, -0.1364789605140686, -0.001243848935700953, -0.20151029527187347, 0.03967966139316559, 0.013079567812383175, -0.0236991997808218, 0.018869584426283836, 0.01450052298605442, -0.0015552955446764827, -0.028510641306638718, 0.006941767875105143, -0.0046331314370036125, -0.026541125029325485, -0.06940009444952011, -0.003802439197897911, -0.11050795018672943, -0.010042325593531132, -0.060664571821689606, 0.04135625436902046, -0.09718462079763412, -0.05409298464655876, 0.045907698571681976, 0.04011012613773346, -0.031095363199710846, -0.039515845477581024, 0.01961369998753071, 0.08837220072746277, -0.04052206128835678, -0.05257798358798027, 0.021400365978479385, 0.017368776723742485, 0.0019796141423285007, 0.08427518606185913, -0.03436272591352463, -0.07285477221012115, 0.09676862508058548, -0.011110194958746433, -0.01958780363202095, 0.04209623485803604, -0.03472542017698288, 0.0513153076171875, -0.005000270903110504, 0.00829863827675581, 0.2228686660528183, 0.016078131273388863, 0.10014010965824127, -0.1865101307630539, -0.04571036994457245, 0.014848248101770878, -0.019433844834566116, -0.09603407233953476, 0.16396020352840424, -0.0029502147808670998, -0.15557096898555756, 0.08574088662862778, -0.001925811404362321, 0.0026355343870818615, 0.1933518797159195, -0.00790537428110838, -0.09629027545452118, 0.02487769164144993, 0.009209154173731804, 0.018224159255623817, 0.10632622241973877, -0.11470167338848114, -0.024394869804382324, 0.026900731027126312, 0.000465965858893469, 0.10687901824712753, -0.08553361892700195, 0.03163570910692215, -0.010958072729408741, -0.052172284573316574, -0.04410422220826149, 0.03157183900475502, -0.018234536051750183, 0.09024451673030853, 0.08624640852212906, -0.023139650002121925, -0.06763682514429092, -0.04168030619621277, -0.08218017965555191, 0.19425645470619202, -0.11789576709270477, -0.24171994626522064, -0.1797557920217514, 0.04641985148191452, -0.023926211521029472, 0.10135611146688461, -0.0064689358696341515, -0.08134932816028595, -0.025383686646819115, 0.026379019021987915, 0.14098665118217468, -0.07433120906352997, -0.09776702523231506, -0.04750089347362518, 0.02673872746527195, -0.07737230509519577, -0.1677597612142563, 0.028570665046572685, -0.02383122220635414, -0.09555168449878693, -0.03020313009619713, -0.07675998657941818, 0.12471578270196915, 0.08939588069915771, 0.012006347067654133, -0.027042394503951073, -0.026643965393304825, 0.15883323550224304, -0.15405778586864471, -0.03499139845371246, 0.1601577252149582, 0.0935959666967392, -0.03625234588980675, 0.03062358684837818, -0.01811346970498562, -0.05165449529886246, -0.03597148880362511, 0.02969193086028099, -0.05733885616064072, -0.2739703953266144, -0.11674144864082336, -0.023866791278123856, -0.08348827064037323, 0.0611875094473362, 0.024301547557115555, -0.012058043852448463, 0.09459259361028671, -0.02918068692088127, -0.05245376005768776, -0.03351191058754921, 0.04434559866786003, 0.11957215517759323, 0.011146420612931252, 0.03890642523765564, -0.07358258962631226, -0.035441771149635315, 0.15076640248298645, 0.05368681252002716, 0.08779226988554001, -0.03930119052529335, 0.11103616654872894, 0.0866648480296135, 0.07161954045295715, -0.02096252143383026, 0.0652826800942421, 0.007159861270338297, 0.03097916953265667, -0.041554052382707596, -0.05444210395216942, -0.11458784341812134, -0.046884264796972275, -0.022719575092196465, 0.012496759183704853, -0.10235897451639175, -0.1496293991804123, 0.024844203144311905, 0.15073059499263763, 0.03444444760680199, -0.23169323801994324, -0.06390707939863205, 0.045286525040864944, -0.02778336964547634, -0.1050754189491272, 0.002842184854671359, -0.0007512380834668875, -0.12616963684558868, 0.08877597004175186, 0.020069049671292305, 0.08840816468000412, -0.11493739485740662, -0.016697702929377556, -0.08022499084472656, 0.020884912461042404, 0.0012101257452741265, 0.08575539290904999, -0.08955162763595581, 0.23568759858608246, 0.027109725400805473, 0.06870165467262268, -0.028890950605273247, 0.03320295736193657, -0.028601547703146935, -0.0014087097952142358, 0.14609842002391815, 0.015232701785862446, -0.013963202945888042, -0.032135944813489914, -0.03290809690952301, 0.04522143676877022, -0.024690883234143257, -0.10414620488882065, 0.10412255674600601, 0.034731969237327576, 0.004814486484974623, -0.08039790391921997, -0.08557526022195816, -0.03415033966302872, -0.09903534501791, 0.012898780405521393, -0.0828808844089508, 0.006169628351926804, -0.04204780235886574, -0.04988069832324982, -0.049112364649772644, 0.11582610756158829, -0.0593150369822979, -0.08218173682689667, -0.12238213419914246, 0.07383906841278076, 0.15568582713603973, -0.06707026809453964, 0.013657069765031338, -0.053260717540979385, 0.04917630925774574, -0.0197968278080225, -0.05550196394324303, 0.053617291152477264, -0.11798890680074692, -0.04523749276995659, -0.01586204394698143, 0.07163398712873459, 0.0842585563659668, -0.00010507582919672132, 0.06830262392759323, 0.03688056021928787, 0.008931166492402554, -0.07606510072946548, -0.010424982756376266, -0.05442173406481743, 0.0785706415772438, 0.07817953079938889, -0.04531016945838928, -0.18497446179389954, -0.03937866911292076, 0.013630562461912632, 0.04910624772310257, 0.07868800312280655, -0.007073056884109974, 0.07515367120504379, 0.24520595371723175, -0.12484101206064224, -0.1666073203086853, 0.015817780047655106, 0.0900939404964447, 0.04655788093805313, -0.06718679517507553, -0.23514431715011597, 0.030126335099339485, 0.1587206870317459, -0.03797474130988121, 0.023300034925341606, -0.31395524740219116, -0.02490374818444252, 0.10573296993970871, 0.06663161516189575, 0.09800752252340317, -0.07755730301141739, -0.0974401980638504, -0.01578669250011444, -0.027270808815956116, 0.045337431132793427, -0.007357222028076649, 0.02349584363400936, -0.012155788019299507, -0.058509889990091324, 0.0478312149643898, 0.0008286047377623618, 0.14893370866775513, -0.02411591075360775, 0.021729709580540657, 0.009723126888275146, 0.06132614240050316, 0.14649517834186554, 0.03022707626223564, 0.06679581850767136, 0.16766344010829926, 0.030632292851805687, -0.20689845085144043, -0.04124992713332176, -0.07672654092311859, 0.07796648144721985, -0.058403078466653824, -0.002088781213387847, -0.03064683824777603, 0.0596439354121685, 0.01744619570672512, -0.00463350722566247, 0.07216047495603561, -0.07477067410945892, 0.11393333971500397, 0.10096606612205505, 0.1674344390630722, 0.08507297188043594, 0.008722382597625256, 0.03861599043011665, 0.013237828388810158, 0.05648159980773926, -0.01201460137963295, 0.0029850872233510017, 0.10906145721673965, 0.007802370935678482, -0.01572604291141033, 0.0028031407855451107, -0.17000912129878998, 0.03508731350302696, 0.16636335849761963, 0.00006953824049560353, 0.02010459639132023, -0.045604486018419266, -0.08272632956504822, -0.0004538849461823702, 0.006207134108990431, 0.14939849078655243, 0.04321742057800293, -0.09097999334335327, -0.015381556004285812, -0.003849948523566127, -0.006782961077988148, 0.13428793847560883, -0.00889364443719387, 0.013006684370338917, -0.09984248131513596, 0.10953377932310104, 0.11060252040624619, -0.11186639219522476, -0.009055566973984241, 0.07048460096120834, -0.027623271569609642, -0.040248993784189224, 0.006779287476092577, 0.08912355452775955, -0.2735616862773895, -0.09991173446178436, -0.09999959170818329, -0.08009139448404312, 0.0018019110430032015, 0.20063909888267517, 0.04014543071389198, 0.033559512346982956, 0.023870820179581642, 0.012741658836603165, -0.01350882463157177, -0.018204817548394203, 0.029932498931884766, -0.016439031809568405, 0.010111942887306213, 0.05933908745646477, 0.01226096972823143, -0.04157847538590431, -0.026862556114792824, -0.01988920196890831, -0.10236592590808868, 0.01732516847550869, -0.09576576948165894, 0.01961081475019455, -0.05128490552306175, -0.02009897492825985, -0.04280368983745575, -0.001424708985723555, -0.03716585040092468, -0.0394408218562603, -0.06122671440243721, 0.011816742829978466, -0.019857918843626976, 0.08029238134622574, -0.10038799047470093, 0.0349268838763237, -0.00914048682898283, -0.036286626011133194, 0.01991092599928379, 0.017975518479943275, 0.059871502220630646, 0.11458444595336914, -0.12125683575868607, 0.04719555750489235, 0.04052383825182915, 0.021159173920750618, 0.04766124114394188, -0.05104517564177513, 0.053976502269506454, 0.08503177762031555, -0.020896095782518387, 0.04417853057384491, -0.03746408596634865, -0.11103673279285431, -0.02709660679101944, 0.06310085207223892, -0.04773500934243202, -0.01330170501023531, 0.13825669884681702, 0.16264425218105316, 0.04990522935986519, 0.05487823486328125, -0.06735900789499283, -0.015526455827057362, -0.10941530019044876, -0.013997448608279228, -0.04802730679512024, 0.027075257152318954, -0.057795874774456024, 0.029483553022146225, 0.040275923907756805, 0.03683377057313919, 0.15072505176067352, 0.08749744296073914, 0.17215296626091003, 0.021713541820645332, -0.01151980459690094, -0.020326605066657066, 0.011835084296762943, 0.1311175525188446, 0.11187704652547836, 0.019662605598568916, 0.030170511454343796, 0.08836963027715683, -0.019175034016370773, -0.05128342658281326, 0.039118003100156784, 0.17198114097118378, 0.20917493104934692, 0.11270247399806976, 0.03379151225090027, -0.024738740175962448, 0.04475540667772293, 0.033677179366350174, 0.005194083787500858, -0.0028870885726064444, 0.04370562732219696, 0.012696114368736744, 0.11961359530687332, -0.14737896621227264, 0.09251012653112411, 0.06279369443655014, 0.006029644515365362, -0.085405133664608, -0.12859150767326355, -0.009023590944707394, -0.05265232175588608, 0.015223975293338299, -0.13596667349338531, 0.027542468160390854, -0.04946628212928772, 0.021736990660429, -0.09022834897041321, 0.02391825057566166, -0.058866869658231735, -0.18094538152217865, 0.11666339635848999, 0.03764541447162628, 0.18843553960323334, -0.014279681257903576, 0.06000763177871704, 0.011951977387070656, 0.16476289927959442, 0.04915718734264374, 0.05842214450240135, 0.011268776841461658, 0.05629827454686165, -0.025067303329706192, -0.038474421948194504, 0.01030732411891222, 0.04528377950191498, 0.07334499061107635, 0.19906508922576904, 0.03895767033100128, -0.027901865541934967, 0.0007890116539783776, 0.23827244341373444, -0.014064493589103222, 0.06000867113471031, -0.09608284384012222, 0.20612932741641998, 0.039644576609134674, 0.04844871908426285, 0.02701367437839508, -0.13216155767440796, -0.03549445420503616, 0.16162635385990143, 0.12239707261323929, 0.06291181594133377, -0.04961712285876274, 0.011950169689953327, 0.001997275510802865, 0.0762169137597084, 0.06417789310216904, -0.015956353396177292, 0.3456013798713684, -0.05508662015199661, -0.015913791954517365, -0.05836433172225952, 0.10167615115642548, 0.005867823492735624, 0.12100812792778015, -0.04978340491652489, -0.06443009525537491, -0.017197443172335625, 0.16580359637737274, -0.08733910322189331, -0.30626416206359863, 0.06865454465150833, -0.09479489177465439, -0.15617883205413818, -0.05830376595258713, -0.02125023864209652, 0.08095389604568481, 0.08168948441743851, 0.06418652832508087, -0.04501454532146454, 0.04181503877043724, 0.04457058385014534, -0.08085417002439499, -0.2158319354057312, 0.0782589241862297, 0.006292100064456463, 0.26687994599342346, -0.024342559278011322, -0.002250897465273738, 0.07940711081027985, 0.02124350517988205, -0.09921477735042572, 0.03306305408477783, 0.01888713613152504, -0.015241007320582867, 0.10301661491394043, 0.015364410355687141, 0.030832568183541298, 0.031982507556676865, 0.06219407543540001, -0.0749773308634758, 0.04864073544740677, 0.024957938119769096, -0.01504316832870245, -0.11648894101381302, 0.10885051637887955, -0.12769635021686554, 0.0794009417295456, 0.17782732844352722, 0.003760596737265587, 0.035188425332307816, -0.06871432811021805, 0.09558505564928055, -0.003964447882026434, 0.13168896734714508, -0.01026900950819254, -0.16185756027698517, -0.011214343830943108, -0.04505334421992302, -0.04177156835794449, -0.22148218750953674, 0.10887986421585083, 0.03462766110897064, -0.04665501043200493, -0.01578523963689804, 0.0370035246014595, -0.08439771831035614, 0.05608503147959709, 0.009431350976228714, 0.031258728355169296, 0.0022323220036923885, 0.05025804415345192, -0.08581553399562836, -0.030163930729031563 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/roberta-base-bne-capitel-pos # Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Part of Speech (POS) dataset RoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019. Original pre-trained model can be found here: https://huggingface.co/BSC-TeMU/roberta-base-bne ## Dataset The dataset used is the one from the [CAPITEL competition at IberLEF 2020](https://sites.google.com/view/capitel2020) (sub-task 2). ## Evaluation and results F1 Score: 0.9846 (average of 5 runs). For evaluation details visit our [GitHub repository](https://github.com/PlanTL-SANIDAD/lm-spanish). ## Citing Check out our paper for all the details: https://arxiv.org/abs/2107.07253 ``` @misc{gutierrezfandino2021spanish, title={Spanish Language Models}, author={Asier Gutiérrez-Fandiño and Jordi Armengol-Estapé and Marc Pàmies and Joan Llop-Palao and Joaquín Silveira-Ocampo and Casimiro Pio Carrino and Aitor Gonzalez-Agirre and Carme Armentano-Oller and Carlos Rodriguez-Penagos and Marta Villegas}, year={2021}, eprint={2107.07253}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
{"language": ["es"], "license": "apache-2.0", "tags": ["national library of spain", "spanish", "bne", "capitel", "pos"], "datasets": ["bne", "capitel"], "metrics": ["f1"], "widget": [{"text": "Festival de San Sebasti\u00e1n: Johnny Depp recibir\u00e1 el premio Donostia en pleno rifirrafe judicial con Amber Heard"}, {"text": "El alcalde de Vigo, Abel Caballero, ha comenzado a colocar las luces de Navidad en agosto."}, {"text": "Gracias a los datos de la BNE, se ha podido lograr este modelo del lenguaje."}, {"text": "El Tribunal Superior de Justicia se pronunci\u00f3 ayer: \"Hay base legal dentro del marco jur\u00eddico actual\"."}]}
token-classification
BSC-LT/roberta-base-bne-capitel-pos
[ "transformers", "pytorch", "roberta", "token-classification", "national library of spain", "spanish", "bne", "capitel", "pos", "es", "dataset:bne", "dataset:capitel", "arxiv:1907.11692", "arxiv:2107.07253", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1907.11692", "2107.07253" ]
[ "es" ]
TAGS #transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #pos #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL # Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Part of Speech (POS) dataset RoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019. Original pre-trained model can be found here: URL ## Dataset The dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 2). ## Evaluation and results F1 Score: 0.9846 (average of 5 runs). For evaluation details visit our GitHub repository. ## Citing Check out our paper for all the details: URL
[ "# Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Part of Speech (POS) dataset\nRoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 2).", "## Evaluation and results\nF1 Score: 0.9846 (average of 5 runs).\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ "TAGS\n#transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #pos #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Part of Speech (POS) dataset\nRoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 2).", "## Evaluation and results\nF1 Score: 0.9846 (average of 5 runs).\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ 95, 137, 29, 32, 13 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #pos #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Spanish RoBERTa-base trained on BNE finetuned for CAPITEL Part of Speech (POS) dataset\nRoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 2).## Evaluation and results\nF1 Score: 0.9846 (average of 5 runs).\n\nFor evaluation details visit our GitHub repository.## Citing \nCheck out our paper for all the details: URL" ]
[ -0.13387423753738403, 0.24093370139598846, -0.004012162331491709, 0.019911877810955048, 0.054934341460466385, -0.026720626279711723, 0.023411985486745834, 0.10841643810272217, -0.02666184864938259, 0.09295381605625153, 0.01703743450343609, 0.08128944039344788, 0.09577439725399017, 0.08500874042510986, 0.019009394571185112, -0.19769208133220673, 0.058640703558921814, -0.024968890473246574, 0.028007017448544502, 0.06164143607020378, 0.09587322920560837, -0.05264702066779137, 0.0037302840501070023, 0.005098635330796242, -0.05034502595663071, 0.09272750467061996, -0.0283272136002779, -0.15607614815235138, 0.07146493345499039, 0.024208949878811836, 0.10314953327178955, 0.02598792314529419, 0.01855894923210144, -0.031084703281521797, -0.0011727516539394855, 0.047556422650814056, 0.006941435858607292, 0.0440157987177372, 0.1281082034111023, -0.1382809579372406, 0.20080742239952087, 0.008003159426152706, 0.015633726492524147, 0.046662840992212296, -0.18788036704063416, -0.1937514692544937, -0.1254410594701767, -0.016737980768084526, 0.06602955609560013, 0.10963629186153412, -0.02713124454021454, 0.08592025190591812, -0.02279467135667801, -0.02418588288128376, 0.08294940739870071, -0.18921254575252533, -0.06390395760536194, 0.09094183146953583, 0.039578069001436234, 0.089195117354393, -0.053461410105228424, 0.00890783779323101, 0.08906503021717072, 0.011759716086089611, -0.045912034809589386, -0.05882607400417328, -0.12465745955705643, -0.021373677998781204, -0.12583981454372406, -0.10173448175191879, 0.10025043040513992, -0.08583933115005493, -0.07023391872644424, -0.02312535233795643, -0.06325680017471313, -0.002534240484237671, 0.01445799320936203, 0.0032014900352805853, 0.0671033039689064, 0.009474574588239193, 0.034800611436367035, -0.05677386373281479, -0.06934057176113129, -0.022266291081905365, -0.08598669618368149, 0.19193971157073975, 0.014822524040937424, 0.026043709367513657, 0.019929388538002968, 0.0819779634475708, -0.007160489913076162, -0.08982790261507034, -0.06190364062786102, -0.001455347053706646, -0.0250746700912714, -0.02850295789539814, -0.05038461461663246, -0.13434220850467682, 0.024045461788773537, 0.10815512388944626, -0.1851731687784195, -0.019212201237678528, -0.0369609072804451, 0.005058008711785078, 0.10673611611127853, 0.1567232608795166, -0.04718392714858055, -0.0752885565161705, 0.0044542476534843445, -0.05137750133872032, 0.011181192472577095, 0.03388999029994011, -0.04245161637663841, -0.06502280384302139, -0.01830742321908474, 0.08656363189220428, 0.06051017343997955, -0.06235896050930023, -0.04134242609143257, -0.04664193093776703, 0.13879847526550293, -0.0893721878528595, -0.04170352965593338, -0.037076447159051895, -0.1064196228981018, 0.01566619612276554, -0.023290831595659256, -0.03002898581326008, -0.0866500735282898, -0.06773000210523605, -0.07514001429080963, -0.10874335467815399, -0.07473980635404587, -0.07813133299350739, 0.06746801733970642, -0.18009400367736816, -0.009463093243539333, -0.0881817638874054, -0.09111297130584717, -0.12817837297916412, 0.014543801546096802, -0.09183521568775177, 0.04539349302649498, -0.052891235798597336, 0.06594107300043106, 0.0010069360723719, -0.024143097922205925, 0.003455393947660923, -0.03491032496094704, 0.05319923162460327, 0.038104187697172165, 0.06573084741830826, -0.06382155418395996, 0.004936905577778816, -0.16462083160877228, -0.003648454090580344, -0.2524244785308838, 0.1134544238448143, -0.07694949209690094, 0.02194412611424923, -0.17218580842018127, -0.029977353289723396, -0.04988812282681465, 0.07662884891033173, 0.07167373597621918, 0.122923344373703, -0.04039863124489784, -0.07932201772928238, 0.21366828680038452, -0.11586609482765198, -0.03561769425868988, 0.060373615473508835, 0.016391867771744728, 0.09164085239171982, 0.10749530792236328, 0.11070011556148529, 0.0481484979391098, -0.05294634774327278, -0.09291429817676544, -0.034157223999500275, 0.05688156187534332, -0.10826965421438217, 0.09652885794639587, -0.12267521768808365, 0.010171118192374706, 0.011164389550685883, -0.13882893323898315, -0.005769224371761084, -0.002941869432106614, -0.019673483446240425, 0.0364513024687767, -0.011164332740008831, -0.099355548620224, 0.04318780452013016, 0.035417698323726654, -0.01894816942512989, -0.038970839232206345, 0.019540123641490936, 0.04566894471645355, 0.015514452941715717, 0.0025062572676688433, -0.036600541323423386, 0.11661732196807861, -0.13190414011478424, -0.005873701069504023, -0.18159660696983337, 0.029256997630000114, 0.00009714795305626467, -0.02757834829390049, 0.045850999653339386, 0.00010232498607365415, 0.006666500121355057, -0.01575433276593685, -0.013189953751862049, -0.012051603756844997, -0.005287060514092445, -0.06859384477138519, -0.01960046961903572, -0.1042119488120079, -0.0015565428184345365, -0.06265580654144287, 0.03783253952860832, -0.07838909327983856, -0.044542714953422546, 0.021843839436769485, 0.0945257842540741, -0.031428609043359756, -0.0625806525349617, 0.05165217071771622, 0.05283205211162567, -0.02532334439456463, -0.08464541286230087, 0.007216820493340492, 0.03448450192809105, 0.02516462467610836, 0.09805967658758163, -0.02825477160513401, -0.09168415516614914, 0.1043996587395668, 0.030263669788837433, -0.019014589488506317, 0.028032422065734863, -0.04023327678442001, 0.056511711329221725, -0.0461324080824852, 0.017191480845212936, 0.20649394392967224, 0.0023366552777588367, 0.09465820342302322, -0.16692830622196198, -0.011017350479960442, 0.0069559793919324875, -0.02245865948498249, -0.08290757238864899, 0.12891988456249237, 0.012761669233441353, -0.1541728526353836, 0.07934534549713135, -0.04333942010998726, -0.0005002935649827123, 0.22526238858699799, 0.010026215575635433, -0.09641413390636444, 0.021103307604789734, 0.008901050314307213, 0.015158604830503464, 0.11040759086608887, -0.13237006962299347, -0.042434561997652054, 0.0069076200015842915, 0.0085531584918499, 0.10519888997077942, -0.08797914534807205, 0.02733840048313141, -0.018778249621391296, -0.047876205295324326, 0.0016830554232001305, 0.02528338134288788, -0.03382260352373123, 0.08470393717288971, 0.0947631224989891, -0.023237871006131172, -0.06480934470891953, -0.021232416853308678, -0.06699614971876144, 0.18590371310710907, -0.11454514414072037, -0.22563359141349792, -0.1649577021598816, 0.031019674614071846, -0.038319651037454605, 0.10689537227153778, 0.011211499571800232, -0.09700679779052734, -0.0374026782810688, 0.026506846770644188, 0.16139720380306244, -0.07640179246664047, -0.07375969737768173, -0.07566087692975998, 0.02168240025639534, -0.06943055242300034, -0.16633997857570648, 0.019320251420140266, -0.011992160230875015, -0.1355678290128708, -0.03313605114817619, -0.10231202095746994, 0.07764172554016113, 0.09062390774488449, 0.03765023872256279, -0.00397771131247282, -0.027892353013157845, 0.16533368825912476, -0.14515097439289093, -0.03010553866624832, 0.16049900650978088, 0.12360166013240814, -0.038029998540878296, 0.06323646008968353, -0.01253784541040659, -0.08683587610721588, -0.028943412005901337, 0.04228564724326134, -0.06800279766321182, -0.2758651673793793, -0.11103511601686478, -0.033578984439373016, -0.09436783194541931, 0.10898709297180176, 0.0421699658036232, 0.014380337670445442, 0.09823056310415268, -0.07348477095365524, -0.022918350994586945, -0.03765686973929405, 0.06585510820150375, 0.04049506410956383, 0.00915240217000246, 0.03853408619761467, -0.050194498151540756, -0.00952170416712761, 0.13668596744537354, 0.06815774738788605, 0.07340825349092484, -0.03947578743100166, 0.10355465114116669, 0.08829335123300552, 0.07336634397506714, -0.02037026733160019, 0.057577945291996, 0.011181301437318325, 0.03342226520180702, -0.029905900359153748, -0.05621834099292755, -0.11621950566768646, -0.009339782409369946, -0.014375409111380577, -0.027323871850967407, -0.0762895867228508, -0.16151097416877747, 0.006656917277723551, 0.1954403966665268, 0.06862878054380417, -0.2272646725177765, -0.07362812757492065, 0.04160716012120247, -0.03998107463121414, -0.11985132843255997, 0.03512483462691307, -0.013130669482052326, -0.09588402509689331, 0.07416021078824997, 0.017194688320159912, 0.0830620601773262, -0.134914830327034, -0.020450612530112267, -0.07175678014755249, 0.04894844815135002, -0.012725690379738808, 0.060075365006923676, -0.09928177297115326, 0.22583961486816406, 0.029404209926724434, 0.07850766181945801, -0.03237854316830635, 0.03482529893517494, -0.009249594993889332, -0.028568614274263382, 0.12729807198047638, 0.030649352818727493, -0.03339831531047821, -0.021527763456106186, -0.04446396604180336, 0.03832409158349037, -0.020594624802470207, -0.12863337993621826, 0.10470333695411682, 0.03767233341932297, -0.003939344547688961, -0.0806809812784195, -0.10154710710048676, -0.06853815168142319, -0.0882643610239029, 0.030507894232869148, -0.06043523550033569, 0.07311937212944031, -0.05503759905695915, -0.03687242045998573, -0.009354081004858017, 0.15474744141101837, -0.0022840341553092003, -0.0827052965760231, -0.13032756745815277, 0.0854361355304718, 0.15228143334388733, -0.044261634349823, 0.030405020341277122, -0.045092761516571045, 0.06691741943359375, -0.0216389037668705, -0.05061330273747444, 0.0735173299908638, -0.12987571954727173, -0.05455129221081734, -0.018547870218753815, 0.07394774258136749, 0.09476020932197571, -0.003824685700237751, 0.06094782054424286, 0.05136943235993385, 0.010114475153386593, -0.06018996983766556, 0.00618106359615922, -0.010322267189621925, 0.08439507335424423, 0.08607441186904907, -0.016904238611459732, -0.13749751448631287, -0.045095913112163544, 0.0003341476258356124, 0.01045843306928873, 0.0453936867415905, 0.01677701622247696, 0.07073155045509338, 0.24859033524990082, -0.11977938562631607, -0.15892422199249268, 0.007992468774318695, 0.0780196413397789, 0.04745005443692207, -0.1266074925661087, -0.2736368775367737, 0.04576927423477173, 0.12632101774215698, -0.0402253121137619, -0.02405661530792713, -0.3265223503112793, -0.03864320367574692, 0.09284486621618271, 0.0555456168949604, 0.14092899858951569, -0.08239906281232834, -0.09332679957151413, -0.019762124866247177, -0.006628607865422964, 0.060540251433849335, -0.06308583915233612, 0.04604765400290489, -0.007778903469443321, -0.055565331131219864, 0.046213872730731964, -0.010934856720268726, 0.14492924511432648, -0.03475657105445862, -0.010314591228961945, 0.013864678330719471, 0.09245146811008453, 0.13123716413974762, -0.003733696648851037, 0.07201803475618362, 0.15302400290966034, 0.035947807133197784, -0.20734632015228271, -0.05234364792704582, -0.0684717670083046, 0.07011421769857407, -0.038641639053821564, -0.023873692378401756, -0.02359164133667946, 0.07290912419557571, 0.0258659478276968, 0.008105870336294174, 0.09577742964029312, -0.07860074192285538, 0.14307890832424164, 0.12298764288425446, 0.12433332949876785, 0.038091544061899185, -0.006450333166867495, 0.03545123711228371, 0.006782975047826767, 0.07617289572954178, -0.05765657126903534, -0.011323895305395126, 0.10089374333620071, 0.020990610122680664, -0.006406825035810471, -0.01081828959286213, -0.17584773898124695, 0.02711385115981102, 0.14636112749576569, -0.02933507040143013, 0.03003968670964241, -0.04009557515382767, -0.031551484018564224, 0.00798011478036642, 0.008517878130078316, 0.19121190905570984, 0.020722821354866028, -0.0907026082277298, -0.0198186244815588, 0.0030359551310539246, -0.015695173293352127, 0.1390100121498108, -0.03585505858063698, 0.005924588069319725, -0.11813788115978241, 0.12172018736600876, 0.11645582318305969, -0.09245148301124573, 0.0018477601697668433, 0.07954128086566925, -0.03720429539680481, -0.04500170424580574, -0.004944209475070238, 0.03235923498868942, -0.2637636065483093, -0.07503890991210938, -0.08472009003162384, -0.07775014638900757, 0.007057324983179569, 0.18801143765449524, 0.017235510051250458, 0.031171174719929695, 0.010417487472295761, 0.028154995292425156, -0.020129308104515076, -0.04590633884072304, 0.046561628580093384, 0.0003765269066207111, 0.003674111096188426, 0.06145268306136131, 0.018127337098121643, -0.05899755284190178, -0.020590996369719505, -0.029723823070526123, -0.12135066092014313, 0.02071688324213028, -0.13273745775222778, 0.000710304535459727, -0.06283402442932129, -0.027838511392474174, -0.04825560003519058, -0.02164176106452942, -0.0371130146086216, -0.03924529626965523, -0.06651805341243744, 0.014255872927606106, -0.04616105556488037, 0.07338178902864456, -0.09813356399536133, 0.03794214874505997, -0.001482963329181075, -0.06433981657028198, 0.032568592578172684, 0.04530010744929314, 0.05400630086660385, 0.09109923988580704, -0.08428797870874405, 0.035677455365657806, 0.04216849058866501, 0.002928304485976696, 0.03647814691066742, -0.035245250910520554, 0.04594661295413971, 0.07424268871545792, -0.019229158759117126, 0.03549991548061371, -0.04340765252709389, -0.10448262095451355, -0.03358517959713936, 0.08165410906076431, -0.01614377833902836, -0.014360103756189346, 0.11714164912700653, 0.20336468517780304, 0.039596669375896454, 0.03409341350197792, -0.065597765147686, -0.02729669399559498, -0.10857725143432617, -0.005164738744497299, -0.05354655906558037, 0.0064408243633806705, -0.052919089794158936, 0.02162991091609001, 0.04335156828165054, 0.04060608148574829, 0.16406655311584473, 0.08206953853368759, 0.1618504375219345, 0.027338732033967972, 0.029537171125411987, 0.015365746803581715, 0.013332144357264042, 0.10726474970579147, 0.1194990873336792, 0.029262179508805275, 0.017104510217905045, 0.08502267301082611, 0.016728004440665245, -0.038626447319984436, 0.048846285790205, 0.13829059898853302, 0.24670714139938354, 0.11715180426836014, 0.01584024354815483, -0.045072343200445175, 0.01490164827555418, 0.03936203196644783, 0.003939244896173477, 0.007068943232297897, 0.03281276300549507, -0.02190100960433483, 0.14322614669799805, -0.14833775162696838, 0.07165030390024185, 0.06017870828509331, 0.006290932185947895, -0.08272835612297058, -0.1426081508398056, 0.0009220296051353216, -0.08484157174825668, 0.011695479042828083, -0.1304909586906433, 0.053965505212545395, -0.008996912278234959, 0.04789308086037636, -0.04775731638073921, 0.022667203098535538, -0.011049294844269753, -0.17819125950336456, 0.10100635886192322, 0.04284872114658356, 0.18933168053627014, -0.0026260451413691044, 0.02171083353459835, 0.011618146672844887, 0.13798809051513672, 0.0478772409260273, 0.052106425166130066, 0.027627479285001755, 0.05318029224872589, -0.015075446106493473, -0.03756526857614517, 0.0012162108905613422, 0.03283006697893143, 0.045316439121961594, 0.2089744359254837, 0.040253400802612305, -0.026912108063697815, 0.015692178159952164, 0.26991307735443115, -0.032857395708560944, 0.04197196662425995, -0.12171144038438797, 0.2043452113866806, 0.0362936332821846, 0.036023832857608795, 0.008402824401855469, -0.13039247691631317, -0.03806172311306, 0.17197681963443756, 0.10605644434690475, 0.07072366029024124, -0.045708365738391876, 0.018358243629336357, -0.009814734570682049, 0.06636732071638107, 0.05380783602595329, -0.013210526667535305, 0.3323694169521332, -0.05999760702252388, -0.018291151151061058, -0.024935821071267128, 0.05169714242219925, -0.010155562311410904, 0.08520545810461044, -0.05673873424530029, -0.06273163855075836, -0.009744659066200256, 0.16374574601650238, -0.10457325726747513, -0.3077106773853302, 0.07646526396274567, -0.08489685505628586, -0.1446554958820343, -0.0595870204269886, -0.01991131529211998, 0.09693842381238937, 0.10129361599683762, 0.0577937588095665, -0.037377625703811646, 0.04940042644739151, 0.03630461171269417, -0.08660582453012466, -0.23470903933048248, 0.09100540727376938, -0.01846192590892315, 0.2595197260379791, -0.027194863185286522, -0.03574713319540024, 0.07274391502141953, 0.011301319114863873, -0.12101968377828598, 0.010829848237335682, 0.0070050060749053955, -0.054709937423467636, 0.07505572587251663, 0.048298418521881104, -0.005953175015747547, 0.060062967240810394, 0.05421137437224388, -0.07878565043210983, 0.03800257295370102, 0.07097955048084259, 0.00728239631280303, -0.11833865195512772, 0.09356118738651276, -0.12270962446928024, 0.0974230170249939, 0.17003180086612701, -0.013698827475309372, 0.03709395229816437, -0.05317709222435951, 0.10688357800245285, -0.001851963927038014, 0.08606377243995667, -0.005505179986357689, -0.20002005994319916, -0.026209471747279167, -0.015265041030943394, -0.02921304665505886, -0.18487924337387085, 0.09685588628053665, 0.04904498532414436, -0.026820776984095573, -0.022232821211218834, 0.05407360941171646, -0.040255654603242874, 0.050841160118579865, 0.0011718004243448377, 0.037462327629327774, -0.0030885448213666677, 0.057601891458034515, -0.08719736337661743, -0.05008041113615036 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/roberta-base-bne-sqac # Spanish RoBERTa-base trained on BNE finetuned for Spanish Question Answering Corpus (SQAC) dataset. RoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019. Original pre-trained model can be found here: https://huggingface.co/BSC-TeMU/roberta-base-bne ## Dataset The dataset used is the [SQAC corpus](https://huggingface.co/datasets/BSC-TeMU/SQAC). ## Evaluation and results F1 Score: 0.7923 (average of 5 runs). For evaluation details visit our [GitHub repository](https://github.com/PlanTL-SANIDAD/lm-spanish). ## Citing Check out our paper for all the details: https://arxiv.org/abs/2107.07253 ``` @misc{gutierrezfandino2021spanish, title={Spanish Language Models}, author={Asier Gutiérrez-Fandiño and Jordi Armengol-Estapé and Marc Pàmies and Joan Llop-Palao and Joaquín Silveira-Ocampo and Casimiro Pio Carrino and Aitor Gonzalez-Agirre and Carme Armentano-Oller and Carlos Rodriguez-Penagos and Marta Villegas}, year={2021}, eprint={2107.07253}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
{"language": ["es"], "license": "apache-2.0", "tags": ["national library of spain", "spanish", "bne", "qa", "question answering"], "datasets": ["BSC-TeMU/SQAC"], "metrics": ["f1"]}
question-answering
BSC-LT/roberta-base-bne-sqac
[ "transformers", "pytorch", "roberta", "question-answering", "national library of spain", "spanish", "bne", "qa", "question answering", "es", "dataset:BSC-TeMU/SQAC", "arxiv:1907.11692", "arxiv:2107.07253", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1907.11692", "2107.07253" ]
[ "es" ]
TAGS #transformers #pytorch #roberta #question-answering #national library of spain #spanish #bne #qa #question answering #es #dataset-BSC-TeMU/SQAC #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL # Spanish RoBERTa-base trained on BNE finetuned for Spanish Question Answering Corpus (SQAC) dataset. RoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019. Original pre-trained model can be found here: URL ## Dataset The dataset used is the SQAC corpus. ## Evaluation and results F1 Score: 0.7923 (average of 5 runs). For evaluation details visit our GitHub repository. ## Citing Check out our paper for all the details: URL
[ "# Spanish RoBERTa-base trained on BNE finetuned for Spanish Question Answering Corpus (SQAC) dataset.\nRoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the SQAC corpus.", "## Evaluation and results\nF1 Score: 0.7923 (average of 5 runs).\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ "TAGS\n#transformers #pytorch #roberta #question-answering #national library of spain #spanish #bne #qa #question answering #es #dataset-BSC-TeMU/SQAC #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #endpoints_compatible #region-us \n", "# Spanish RoBERTa-base trained on BNE finetuned for Spanish Question Answering Corpus (SQAC) dataset.\nRoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the SQAC corpus.", "## Evaluation and results\nF1 Score: 0.7923 (average of 5 runs).\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ 90, 139, 14, 32, 13 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #question-answering #national library of spain #spanish #bne #qa #question answering #es #dataset-BSC-TeMU/SQAC #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #endpoints_compatible #region-us \n# Spanish RoBERTa-base trained on BNE finetuned for Spanish Question Answering Corpus (SQAC) dataset.\nRoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL## Dataset\nThe dataset used is the SQAC corpus.## Evaluation and results\nF1 Score: 0.7923 (average of 5 runs).\n\nFor evaluation details visit our GitHub repository.## Citing \nCheck out our paper for all the details: URL" ]
[ -0.0983697772026062, 0.19815368950366974, -0.002695739036425948, 0.012601776048541069, 0.06830661743879318, -0.021442102268338203, 0.03315761312842369, 0.10444556176662445, -0.014850894920527935, 0.06272383779287338, 0.01879185251891613, 0.05794951692223549, 0.07198365777730942, 0.029388224706053734, 0.050437215715646744, -0.16756723821163177, 0.041572507470846176, -0.03329828381538391, 0.03615305572748184, 0.07031858712434769, 0.0670749694108963, -0.0595579519867897, 0.013830081559717655, -0.020603708922863007, -0.04804660379886627, 0.08500412106513977, -0.054454028606414795, -0.1270882785320282, 0.09206418693065643, 0.023853087797760963, 0.09225551784038544, 0.012399067170917988, 0.042783331125974655, -0.049799833446741104, 0.012376594357192516, 0.059067245572805405, 0.0038002668879926205, 0.04497939720749855, 0.09848103672266006, -0.09940467774868011, 0.14559859037399292, 0.0023526435252279043, 0.005386491771787405, 0.04631191864609718, -0.18468855321407318, -0.15360766649246216, -0.09969504177570343, -0.005072524305433035, 0.04710780084133148, 0.11967689543962479, -0.03187117353081703, 0.08894374966621399, -0.04662593826651573, -0.01691952534019947, 0.08436036854982376, -0.16168111562728882, -0.0696236863732338, 0.09046202152967453, 0.07192894071340561, 0.11905109882354736, -0.04463875666260719, 0.024916771799325943, 0.08090393990278244, 0.008359460160136223, -0.043489955365657806, -0.07855808734893799, -0.15713763236999512, -0.00888759270310402, -0.11741971969604492, -0.08424226194620132, 0.12182402610778809, -0.08151248842477798, -0.0651061162352562, -0.013793371617794037, -0.07291001081466675, 0.10045831650495529, 0.024490777403116226, 0.025633109733462334, 0.07329951226711273, 0.007753971964120865, -0.004707525949925184, -0.0900108739733696, -0.07631996273994446, -0.041040945798158646, -0.08655289560556412, 0.15278731286525726, 0.010636393912136555, 0.030029168352484703, -0.025593452155590057, 0.08978626132011414, 0.012791130691766739, -0.09976813942193985, -0.053223866969347, -0.016987474635243416, -0.028884684666991234, -0.015163492411375046, -0.044058769941329956, -0.131846085190773, 0.05309448391199112, 0.12281043827533722, -0.11152110993862152, -0.04106317460536957, -0.05123632401227951, 0.015753256157040596, 0.1058184802532196, 0.1655472368001938, -0.08918369561433792, -0.055869583040475845, 0.014977212995290756, -0.043137501925230026, 0.017252644523978233, 0.037456341087818146, -0.07276268303394318, -0.06611625850200653, -0.05762006342411041, 0.07652828842401505, 0.0971749871969223, -0.03532107174396515, 0.011501655913889408, -0.06432829797267914, 0.10134249180555344, -0.07570608705282211, -0.051058437675237656, -0.025509852916002274, -0.08733092993497849, -0.018169555813074112, -0.018696023151278496, -0.0243519339710474, -0.08763498067855835, -0.08929836004972458, -0.08935464173555374, -0.08753100782632828, -0.08685074001550674, -0.0916980504989624, 0.050623368471860886, -0.19037532806396484, 0.0027436059899628162, -0.10193415731191635, -0.13625575602054596, -0.13369521498680115, 0.020427942276000977, -0.10094393044710159, 0.07900205999612808, -0.03494032472372055, 0.08751236647367477, -0.015025242231786251, -0.028256336227059364, 0.041788626462221146, -0.03595433011651039, 0.0655759796500206, 0.017580915242433548, 0.06688257306814194, -0.06263928860425949, 0.0002764596720226109, -0.17814847826957703, -0.0006775768706575036, -0.18936853110790253, 0.12145807594060898, -0.048831287771463394, 0.0646020770072937, -0.16916856169700623, -0.010233019478619099, -0.08324508368968964, 0.0773041769862175, 0.07535254955291748, 0.13448549807071686, -0.03205905482172966, -0.051447972655296326, 0.21709811687469482, -0.07328522950410843, -0.058005645871162415, 0.07439830899238586, -0.006827881559729576, 0.14305874705314636, 0.10300051420927048, 0.1311938464641571, 0.01856393925845623, -0.03494715318083763, -0.06140274927020073, -0.004190480336546898, 0.0752447098493576, -0.08457998186349869, 0.08624623715877533, -0.1319994330406189, 0.023158228024840355, 0.00891007948666811, -0.15994185209274292, 0.0011423485120758414, 0.002388068474829197, -0.02118716947734356, 0.03589621186256409, -0.02296941541135311, -0.11543706059455872, 0.02372150495648384, 0.026498476043343544, 0.01405238825827837, -0.025154901668429375, -0.002925564767792821, 0.048428699374198914, 0.01315977144986391, -0.005272609181702137, -0.02698509767651558, 0.07421041280031204, -0.15487803518772125, 0.009075496345758438, -0.18743756413459778, 0.031310513615608215, -0.0006784814759157598, -0.00037430625525303185, 0.06362157315015793, -0.01383194699883461, 0.01447451300919056, -0.033732086420059204, -0.004547484219074249, -0.004672415088862181, -0.03116408921778202, -0.06735600531101227, -0.022752078250050545, -0.09305272251367569, -0.006211481988430023, -0.06095327436923981, 0.023145651444792747, -0.04840235412120819, -0.026156749576330185, -0.03724091500043869, 0.05844010040163994, -0.028238384053111076, -0.040450792759656906, 0.06309492141008377, 0.03894677758216858, -0.026488184928894043, -0.07062142342329025, 0.0202257689088583, 0.04159210994839668, 0.01945395953953266, 0.07812506705522537, -0.010862615890800953, -0.09151382744312286, 0.09486527740955353, 0.019129063934087753, -0.006775600835680962, 0.025373758748173714, -0.03725544735789299, 0.045464932918548584, -0.06362871080636978, -0.0066314940340816975, 0.22252845764160156, 0.00044580717803910375, 0.08460845798254013, -0.16953663527965546, -0.027581825852394104, -0.004698034375905991, -0.035153865814208984, -0.040375884622335434, 0.10748131573200226, 0.050989970564842224, -0.16354697942733765, 0.07689511030912399, -0.042056605219841, -0.010342409834265709, 0.19604702293872833, 0.015180431306362152, -0.10313427448272705, 0.0340254046022892, -0.006555990315973759, -0.005366595461964607, 0.13180002570152283, -0.13174550235271454, -0.05636615306138992, 0.013309380039572716, -0.0031280939001590014, 0.09690753370523453, -0.0776219442486763, 0.016207268461585045, -0.02397109754383564, -0.04076395928859711, -0.03677557036280632, 0.010395042598247528, -0.04693559557199478, 0.08523108810186386, 0.11280372738838196, 0.03724014014005661, -0.07426127046346664, -0.030072199180722237, -0.08952009677886963, 0.1932954490184784, -0.0917707309126854, -0.21955811977386475, -0.1504218876361847, 0.032182659953832626, -0.02049732767045498, 0.10947024822235107, 0.020398374646902084, -0.11764848232269287, -0.021495899185538292, 0.02359049767255783, 0.1427880823612213, -0.046463195234537125, -0.08596750348806381, -0.06889203190803528, 0.037733905017375946, -0.06931478530168533, -0.15212593972682953, 0.04392813891172409, -0.01818295568227768, -0.11653339117765427, -0.038819532841444016, -0.08961000293493271, 0.12853077054023743, 0.07486511021852493, 0.05364470183849335, -0.01653897762298584, -0.014831659384071827, 0.18499070405960083, -0.13851721584796906, 0.0009905498009175062, 0.17311124503612518, 0.09847034513950348, -0.042024217545986176, 0.09192671626806259, -0.005425914656370878, -0.06805936247110367, -0.01698910817503929, 0.0237846989184618, -0.06654355674982071, -0.27640631794929504, -0.0769401490688324, -0.03165719658136368, -0.10433776676654816, 0.09529688954353333, 0.05439108982682228, -0.00493515282869339, 0.12272907793521881, -0.07462754845619202, -0.021115079522132874, -0.03508589044213295, 0.06932111084461212, -0.0016280324198305607, -0.0009043622994795442, 0.03658731281757355, -0.050322845578193665, -0.026535505428910255, 0.11203023791313171, 0.0946444645524025, 0.10268494486808777, -0.051179975271224976, 0.12539805471897125, 0.0759868323802948, 0.04164756089448929, -0.00967349112033844, 0.07196374237537384, -0.009774264879524708, 0.040145955979824066, -0.043659698218107224, -0.054435599595308304, -0.10324788838624954, -0.0188455767929554, -0.04372085630893707, -0.016453776508569717, -0.03652128577232361, -0.16024677455425262, 0.043368998914957047, 0.21946173906326294, 0.054616816341876984, -0.21298089623451233, -0.08465728908777237, 0.03624017909169197, -0.018993109464645386, -0.11551712453365326, 0.02207619696855545, 0.017689600586891174, -0.10928498953580856, 0.03335893154144287, 0.03242946416139603, 0.12163326144218445, -0.1062716543674469, -0.00876640435308218, -0.03456025943160057, 0.022676080465316772, -0.023064594715833664, 0.07978499680757523, -0.11687072366476059, 0.23181568086147308, 0.03642395883798599, 0.06658178567886353, -0.02726168930530548, 0.028146136552095413, -0.018521277233958244, -0.0340765006840229, 0.14826419949531555, 0.02527093142271042, 0.026929480955004692, 0.0009368005557917058, -0.05349126085639, 0.06150887534022331, -0.03509030491113663, -0.14137676358222961, 0.12357621639966965, 0.01824709214270115, -0.0030272335279732943, -0.0886194258928299, -0.07114909589290619, -0.06328506022691727, -0.1065475270152092, -0.0034747666213661432, -0.08970670402050018, 0.07661445438861847, -0.04256807267665863, -0.006835324224084616, 0.04562300816178322, 0.13173221051692963, -0.05581081286072731, -0.09965118765830994, -0.11605699360370636, 0.06035023182630539, 0.14066271483898163, -0.03531370311975479, 0.012871126644313335, -0.030682427808642387, 0.03740110620856285, -0.015925267711281776, -0.0619070939719677, 0.070095494389534, -0.14450861513614655, -0.027178550139069557, -0.025975318625569344, 0.06978026777505875, 0.08420056104660034, 0.0016189184971153736, 0.06384465843439102, 0.025600634515285492, -0.032950811088085175, -0.06410614401102066, -0.00018897865083999932, 0.02286321297287941, 0.08347270637750626, 0.11112329363822937, -0.06628181785345078, -0.10731802135705948, -0.04759509861469269, 0.0024137927684932947, 0.016156764701008797, 0.00020154856611043215, 0.016035454347729683, 0.06368941068649292, 0.29531732201576233, -0.13848254084587097, -0.17702849209308624, 0.0037120766937732697, 0.03855035826563835, 0.037657007575035095, -0.12234716862440109, -0.2716655731201172, 0.029169077053666115, 0.11289046704769135, -0.027371564880013466, -0.03914619982242584, -0.3252488076686859, -0.035397231578826904, 0.09328265488147736, 0.03300042450428009, 0.15857958793640137, -0.09967049956321716, -0.09038545936346054, -0.027635619044303894, -0.0047254119999706745, 0.0259324349462986, -0.13219207525253296, 0.04190526530146599, -0.0021770084276795387, -0.03704250603914261, 0.04582960531115532, -0.02119181677699089, 0.13466815650463104, -0.01371004804968834, -0.005537642166018486, -0.009333604015409946, 0.08429185301065445, 0.16023683547973633, 0.002098727272823453, 0.07572811096906662, 0.15525665879249573, 0.05510566756129265, -0.16305522620677948, -0.055286407470703125, -0.07765493541955948, 0.03045225702226162, -0.03087666444480419, -0.04521246626973152, -0.03234720230102539, 0.07363424450159073, 0.013395312242209911, 0.010938003659248352, 0.057081565260887146, -0.09571907669305801, 0.14111439883708954, 0.10752386599779129, 0.15951015055179596, 0.046364277601242065, -0.013434745371341705, 0.04879935830831528, 0.0017816605977714062, 0.09092394262552261, -0.04418552294373512, -0.009938621893525124, 0.09785718470811844, 0.028242867439985275, 0.0019977157935500145, 0.00008151127258315682, -0.14962519705295563, 0.028918050229549408, 0.12173767387866974, -0.038125086575746536, -0.02938965894281864, -0.03808310627937317, -0.08736855536699295, -0.05105840787291527, -0.008037187159061432, 0.17342078685760498, 0.03323223814368248, -0.09085413068532944, -0.007906968705356121, 0.013220911845564842, 0.0030770660378038883, 0.135029599070549, -0.012368028983473778, 0.002692952286452055, -0.103645458817482, 0.11973579227924347, 0.11932515352964401, -0.07695881277322769, -0.015263188630342484, 0.10569938272237778, -0.05059649795293808, -0.04834624379873276, 0.0003453957906458527, 0.009788679890334606, -0.2651135325431824, -0.05524301528930664, -0.09808968007564545, -0.05415995791554451, -0.013845314271748066, 0.1513630598783493, 0.007550861220806837, 0.02287021279335022, 0.024204783141613007, 0.04326256737112999, -0.024843884631991386, -0.014657796360552311, 0.0069891889579594135, 0.006559667643159628, -0.006398758850991726, 0.018654396757483482, 0.009132837876677513, -0.028824789449572563, -0.03493625670671463, -0.021440787240862846, -0.1386837512254715, 0.04089822620153427, -0.19272270798683167, 0.017277268692851067, -0.07252808660268784, -0.03820909559726715, -0.05576739460229874, -0.02760942466557026, -0.03504282236099243, -0.045684363692998886, -0.04544072598218918, 0.020006874576210976, -0.033720992505550385, 0.06600909680128098, -0.09321323782205582, 0.04368368163704872, 0.008992980234324932, -0.05024509131908417, 0.05476393550634384, 0.0538407526910305, 0.02333761565387249, 0.06160508468747139, -0.09610139578580856, 0.022743040695786476, 0.023202434182167053, 0.00733035197481513, 0.034043993800878525, -0.016985593363642693, 0.03819771483540535, 0.042553432285785675, -0.028727948665618896, 0.022078007459640503, -0.020318225026130676, -0.08801637589931488, 0.01105727069079876, 0.08396545797586441, -0.004372604191303253, -0.029206808656454086, 0.1088663637638092, 0.16500647366046906, 0.06269428133964539, 0.01644432730972767, -0.07145698368549347, -0.0011479557724669576, -0.13114699721336365, -0.006068893242627382, -0.03990817070007324, 0.025293715298175812, -0.07609899342060089, -0.011631999164819717, 0.04114588722586632, 0.019585024565458298, 0.18105530738830566, 0.08685323596000671, 0.19043968617916107, 0.016127925366163254, 0.044235095381736755, 0.026386668905615807, 0.01305319368839264, 0.11697941273450851, 0.09633703529834747, 0.00560658797621727, -0.020684966817498207, 0.07911573350429535, -0.0110530536621809, -0.030051860958337784, 0.05229445546865463, 0.1554975062608719, 0.30993565917015076, 0.0877353847026825, 0.04909152165055275, -0.04028796777129173, 0.015971459448337555, 0.030143456533551216, 0.04594070836901665, -0.0032347117085009813, -0.00419382331892848, -0.0543183870613575, 0.17742402851581573, -0.13840223848819733, 0.06622324138879776, 0.05459237098693848, 0.011496257036924362, -0.08345407992601395, -0.13025091588497162, 0.002275177277624607, -0.08640559762716293, -0.010648622177541256, -0.1367015838623047, 0.020922360941767693, -0.030577009543776512, 0.02105025015771389, -0.05709807202219963, 0.05342675372958183, 0.021899288520216942, -0.1990266889333725, 0.06626850366592407, 0.04095245897769928, 0.19519132375717163, -0.018831992521882057, 0.03995629400014877, 0.03567080199718475, 0.12537533044815063, 0.05342638120055199, 0.06142308562994003, 0.04043394327163696, 0.05331362038850784, -0.04542379453778267, -0.029146414250135422, -0.005577828269451857, 0.05660055950284004, 0.04321610927581787, 0.18744657933712006, 0.05566590651869774, -0.04975644126534462, 0.02351323328912258, 0.29417330026626587, -0.04113921523094177, 0.029797352850437164, -0.1188335120677948, 0.1664058119058609, 0.004578953143209219, 0.030361386016011238, 0.013195235282182693, -0.12242677062749863, -0.0403105802834034, 0.2175576388835907, 0.14785274863243103, 0.01648055575788021, -0.04595130309462547, 0.016173001378774643, -0.003532454138621688, 0.06600106507539749, 0.055926188826560974, 0.012322614900767803, 0.3699239492416382, -0.06797956675291061, -0.030686642974615097, 0.012021704576909542, 0.028851544484496117, -0.024466466158628464, 0.0766715258359909, -0.04965921491384506, -0.05560838058590889, -0.024336332455277443, 0.16001489758491516, -0.13838577270507812, -0.31487905979156494, 0.035283174365758896, -0.07394970953464508, -0.1326398402452469, -0.0584474541246891, -0.04820948466658592, 0.09632106125354767, 0.09477297216653824, 0.04037304222583771, -0.034772004932165146, 0.05813263729214668, 0.037690795958042145, -0.07873605191707611, -0.2231816202402115, 0.1219792366027832, 0.024402599781751633, 0.2259952276945114, -0.04333168640732765, -0.05224357917904854, 0.06813672930002213, -0.005781090818345547, -0.15196603536605835, -0.011459295637905598, 0.013914075680077076, -0.046742383390665054, 0.08239611983299255, 0.025172073394060135, -0.0029289915692061186, 0.04166409745812416, 0.0753670483827591, -0.08243672549724579, 0.04278157278895378, 0.09522616118192673, 0.031165800988674164, -0.14902085065841675, 0.09447308629751205, -0.12735615670681, 0.11601193994283676, 0.149565190076828, -0.02545315772294998, 0.048474229872226715, -0.04462044686079025, 0.0891391858458519, 0.011535308323800564, 0.09626469761133194, -0.0024512670934200287, -0.15814991295337677, -0.025853829458355904, -0.06915401667356491, -0.029590940102934837, -0.22760775685310364, 0.09049497544765472, 0.06523969024419785, -0.011841220781207085, -0.006772560067474842, 0.04492199048399925, 0.004802383948117495, 0.06833720952272415, -0.008049213327467442, 0.023756347596645355, 0.00028688463498838246, 0.04734343662858009, -0.09417343884706497, -0.07056810706853867 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/roberta-base-bne # RoBERTa base trained with data from National Library of Spain (BNE) ## Model Description RoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019. ## Training corpora and preprocessing The [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) crawls all .es domains once a year. The training corpus consists of 59TB of WARC files from these crawls, carried out from 2009 to 2019. To obtain a high-quality training corpus, the corpus has been preprocessed with a pipeline of operations, including among the others, sentence splitting, language detection, filtering of bad-formed sentences and deduplication of repetitive contents. During the process document boundaries are kept. This resulted into 2TB of Spanish clean corpus. Further global deduplication among the corpus is applied, resulting into 570GB of text. Some of the statistics of the corpus: | Corpora | Number of documents | Number of tokens | Size (GB) | |---------|---------------------|------------------|-----------| | BNE | 201,080,084 | 135,733,450,668 | 570GB | ## Tokenization and pre-training The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original [RoBERTA](https://arxiv.org/abs/1907.11692) model with a vocabulary size of 50,262 tokens. The RoBERTa-base-bne pre-training consists of a masked language model training that follows the approach employed for the RoBERTa base. The training lasted a total of 48 hours with 16 computing nodes each one with 4 NVIDIA V100 GPUs of 16GB VRAM. ## Evaluation and results For evaluation details visit our [GitHub repository](https://github.com/PlanTL-SANIDAD/lm-spanish). ## Citing Check out our paper for all the details: https://arxiv.org/abs/2107.07253 ``` @misc{gutierrezfandino2021spanish, title={Spanish Language Models}, author={Asier Gutiérrez-Fandiño and Jordi Armengol-Estapé and Marc Pàmies and Joan Llop-Palao and Joaquín Silveira-Ocampo and Casimiro Pio Carrino and Aitor Gonzalez-Agirre and Carme Armentano-Oller and Carlos Rodriguez-Penagos and Marta Villegas}, year={2021}, eprint={2107.07253}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
{"language": ["es"], "license": "apache-2.0", "tags": ["national library of spain", "spanish", "bne"], "datasets": ["bne"], "metrics": ["ppl"], "widget": [{"text": "Este a\u00f1o las campanadas de La Sexta las presentar\u00e1 <mask>."}, {"text": "David Broncano es un presentador de La <mask>."}, {"text": "Gracias a los datos de la BNE se ha podido <mask> este modelo del lenguaje."}, {"text": "Hay base legal dentro del marco <mask> actual."}]}
fill-mask
BSC-LT/roberta-base-bne
[ "transformers", "pytorch", "roberta", "fill-mask", "national library of spain", "spanish", "bne", "es", "dataset:bne", "arxiv:1907.11692", "arxiv:2107.07253", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1907.11692", "2107.07253" ]
[ "es" ]
TAGS #transformers #pytorch #roberta #fill-mask #national library of spain #spanish #bne #es #dataset-bne #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL RoBERTa base trained with data from National Library of Spain (BNE) =================================================================== Model Description ----------------- RoBERTa-base-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa base model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019. Training corpora and preprocessing ---------------------------------- The National Library of Spain (Biblioteca Nacional de España) crawls all .es domains once a year. The training corpus consists of 59TB of WARC files from these crawls, carried out from 2009 to 2019. To obtain a high-quality training corpus, the corpus has been preprocessed with a pipeline of operations, including among the others, sentence splitting, language detection, filtering of bad-formed sentences and deduplication of repetitive contents. During the process document boundaries are kept. This resulted into 2TB of Spanish clean corpus. Further global deduplication among the corpus is applied, resulting into 570GB of text. Some of the statistics of the corpus: Tokenization and pre-training ----------------------------- The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original RoBERTA model with a vocabulary size of 50,262 tokens. The RoBERTa-base-bne pre-training consists of a masked language model training that follows the approach employed for the RoBERTa base. The training lasted a total of 48 hours with 16 computing nodes each one with 4 NVIDIA V100 GPUs of 16GB VRAM. Evaluation and results ---------------------- For evaluation details visit our GitHub repository. Citing ------ Check out our paper for all the details: URL
[]
[ "TAGS\n#transformers #pytorch #roberta #fill-mask #national library of spain #spanish #bne #es #dataset-bne #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 83 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #fill-mask #national library of spain #spanish #bne #es #dataset-bne #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.10817181318998337, 0.20310872793197632, -0.006971997208893299, 0.09333059936761856, 0.05503338202834129, 0.022737404331564903, 0.06438587605953217, 0.11308560520410538, 0.02791142277419567, 0.012531583197414875, 0.14112262427806854, 0.22381071746349335, 0.010571506805717945, 0.008289661258459091, -0.08142244815826416, -0.15719783306121826, 0.06974560767412186, 0.01848878152668476, -0.08204543590545654, 0.036870043724775314, 0.08250658214092255, -0.030994903296232224, 0.04517325013875961, -0.05160504952073097, -0.03761931508779526, 0.07400854676961899, 0.014425699599087238, -0.13304610550403595, 0.13990135490894318, 0.06758418679237366, 0.10124371200799942, 0.0445413701236248, -0.027157289907336235, -0.08710289746522903, 0.006310963071882725, -0.04196544736623764, -0.09918443113565445, 0.0781787857413292, 0.018764019012451172, -0.06481882929801941, 0.04297831282019615, 0.03447895124554634, -0.026929007843136787, 0.0074279047548770905, -0.16499505937099457, -0.22537167370319366, -0.10334708541631699, 0.033305153250694275, -0.0057276273146271706, 0.07397481054067612, 0.04957247152924538, 0.10527456551790237, -0.05688602477312088, 0.014960468746721745, 0.16726884245872498, -0.3073540925979614, -0.025846831500530243, -0.01473874319344759, 0.06603731960058212, 0.04028776288032532, 0.015542786568403244, 0.060197364538908005, 0.10865509510040283, -0.018139956519007683, -0.007054226938635111, -0.09804517030715942, -0.15076103806495667, 0.028438420966267586, -0.025940198451280594, -0.08537723124027252, 0.22159872949123383, -0.02874854765832424, 0.03485472872853279, 0.0580519363284111, -0.08136603981256485, 0.0910983458161354, 0.017499569803476334, 0.03539388254284859, 0.034378666430711746, 0.04033819958567619, 0.10044214129447937, -0.0031093547586351633, -0.08808539062738419, 0.020662104710936546, -0.20204627513885498, 0.11696501076221466, 0.0018451330251991749, 0.06272413581609726, -0.03950077295303345, 0.021890787407755852, -0.018595082685351372, -0.1384262591600418, 0.02123132534325123, -0.018524037674069405, 0.122124083340168, 0.06801588833332062, -0.04078782722353935, 0.014328009448945522, 0.11721952259540558, 0.2044961154460907, -0.039703261107206345, -0.032698024064302444, -0.04769380763173103, 0.12680770456790924, 0.013040932826697826, 0.041934024542570114, -0.015285451896488667, -0.07209529727697372, 0.06785821914672852, -0.09601213783025742, 0.07748092710971832, -0.0057055167853832245, -0.17865043878555298, -0.09782926738262177, -0.06658105552196503, 0.12312564998865128, 0.10291008651256561, -0.05367562919855118, -0.06393615156412125, 0.019723132252693176, 0.14587387442588806, -0.034242045134305954, 0.02972959168255329, -0.018756216391921043, -0.03434082865715027, 0.006088219117373228, -0.01841205172240734, -0.028426794335246086, -0.03085458092391491, 0.04406346380710602, -0.08017688244581223, -0.07496082037687302, -0.04340643063187599, -0.054637037217617035, 0.1115397959947586, -0.1217254102230072, 0.07803251594305038, -0.18560706079006195, -0.1579338163137436, 0.010052811354398727, 0.06558765470981598, -0.10760965198278427, -0.012998856604099274, -0.0015155710279941559, 0.02988150529563427, 0.0514078252017498, -0.07960808277130127, 0.022652389481663704, -0.09532810002565384, 0.10252062231302261, -0.005085853394120932, 0.0631510466337204, -0.19648045301437378, -0.0007645636796951294, -0.09707321226596832, 0.01741960644721985, -0.15052801370620728, -0.042944300919771194, -0.09413495659828186, 0.09573666751384735, -0.05946166068315506, -0.004398103803396225, -0.011833026073873043, 0.044620875269174576, 0.05559667572379112, 0.1107974424958229, -0.06946121156215668, -0.07565349340438843, 0.16395121812820435, -0.07688087224960327, -0.1625504046678543, 0.07853519171476364, 0.01990421861410141, 0.048818230628967285, 0.04024643078446388, 0.1737148016691208, -0.009688028134405613, -0.09789152443408966, 0.004176090471446514, 0.07328376919031143, -0.030238337814807892, -0.22561340034008026, 0.09054476022720337, -0.048603519797325134, -0.047133348882198334, 0.035051390528678894, -0.03128790482878685, 0.0635598674416542, -0.00667577376589179, -0.04690471291542053, 0.009684492833912373, -0.05871986970305443, 0.02816617861390114, 0.014038540422916412, 0.0671766921877861, -0.07428208738565445, 0.015555428341031075, -0.014663350768387318, 0.0001450292329536751, 0.07009927183389664, 0.04028483107686043, -0.027946189045906067, 0.1475709229707718, -0.039597898721694946, 0.0006510673556476831, -0.10217699408531189, 0.08661800622940063, -0.01876954734325409, 0.053557686507701874, -0.006174230016767979, 0.016953419893980026, 0.04180486872792244, -0.03926251828670502, -0.05770983546972275, -0.006780494004487991, 0.05839274451136589, 0.02437739446759224, 0.021552179008722305, -0.15763184428215027, 0.0817481279373169, -0.04363637790083885, -0.03930719941854477, -0.04905092343688011, -0.005098492372781038, -0.04902094230055809, 0.06223955750465393, -0.06248205155134201, 0.08961094915866852, -0.07487919926643372, 0.051991160959005356, -0.058300308883190155, -0.002397664589807391, 0.08782599121332169, 0.06030578538775444, -0.03352612629532814, 0.13904452323913574, -0.05833188816905022, 0.283641517162323, 0.16006268560886383, -0.1690659373998642, 0.023541074246168137, 0.007914029061794281, -0.0015134378336369991, 0.006697464268654585, 0.03769538551568985, -0.05052813142538071, 0.049736566841602325, -0.017058832570910454, 0.14538639783859253, -0.12199918925762177, -0.006456729490309954, 0.026665886864066124, -0.067608542740345, -0.06704895198345184, 0.11754345893859863, 0.1411515772342682, -0.12120527774095535, 0.18245482444763184, 0.23870179057121277, -0.05953988805413246, 0.15553289651870728, 0.0010357710998505354, -0.02619016356766224, 0.004115356598049402, -0.05562244728207588, 0.028425397351384163, 0.15218617022037506, -0.10601359605789185, 0.0370376780629158, 0.07006797939538956, -0.006439684424549341, 0.05681126192212105, -0.11265864223241806, -0.07563521713018417, 0.0034206300042569637, 0.0034477682784199715, -0.06157185137271881, 0.05143657699227333, -0.027348371222615242, 0.12242475897073746, 0.0214534904807806, -0.13486899435520172, 0.0690104216337204, 0.016006389632821083, -0.05202067643404007, 0.15921799838542938, -0.16179460287094116, -0.3060198426246643, -0.15914985537528992, -0.0773668885231018, 0.04522779583930969, 0.073092982172966, 0.0984559878706932, -0.043096475303173065, -0.02885107323527336, 0.06682335585355759, -0.017137788236141205, -0.03570364788174629, -0.06102203577756882, -0.01805918663740158, 0.08027146011590958, -0.03371107950806618, -0.13139154016971588, -0.029076414182782173, 0.04379578307271004, 0.011275232769548893, 0.035324886441230774, -0.10119086503982544, 0.1421440988779068, 0.03314097598195076, 0.04505542665719986, 0.010178578086197376, -0.028270365670323372, 0.11974744498729706, -0.04889913648366928, -0.014746599830687046, 0.16818737983703613, 0.02259882539510727, 0.02767498791217804, 0.16357144713401794, 0.038363054394721985, -0.05413886904716492, -0.0508139543235302, -0.05948392674326897, -0.0700814351439476, -0.2888064682483673, -0.11768869310617447, -0.08913568407297134, 0.04416080564260483, 0.06914616376161575, 0.05690670385956764, 0.11944372951984406, 0.08768186718225479, 0.023739423602819443, -0.00831670593470335, -0.0762610137462616, 0.06157086417078972, 0.22159914672374725, -0.012241634540259838, 0.06559343636035919, -0.08014220744371414, -0.049482520669698715, 0.12126415222883224, 0.12145759165287018, 0.05251067131757736, 0.08163748681545258, 0.09041855484247208, 0.024141816422343254, 0.14785702526569366, 0.009941987693309784, 0.10634104162454605, 0.025812938809394836, -0.029366400092840195, -0.0806649848818779, -0.02389652095735073, -0.059643570333719254, -0.013491760939359665, -0.01836290769279003, -0.03462062403559685, -0.05533825606107712, -0.22598662972450256, 0.058078814297914505, 0.045340459793806076, 0.046135421842336655, -0.1624487340450287, 0.020806951448321342, 0.058540452271699905, 0.022043200209736824, -0.10324778407812119, 0.0298799816519022, -0.03175424784421921, -0.11155913025140762, 0.06851822882890701, 0.025316720828413963, 0.12353329360485077, 0.009683790616691113, 0.03306538611650467, -0.1326194703578949, -0.11316040903329849, 0.03122185543179512, 0.1016756072640419, -0.24311187863349915, 0.31350943446159363, 0.023786695674061775, -0.0077596185728907585, -0.04317459464073181, -0.00841534323990345, -0.00013950227003078908, 0.10719068348407745, 0.12066899240016937, 0.02594013325870037, 0.013592536561191082, -0.009584920480847359, -0.03267190605401993, 0.04012689366936684, -0.08158920705318451, -0.02388579770922661, -0.04320641607046127, 0.002234712475910783, -0.01212928257882595, -0.01468789204955101, 0.08467079699039459, -0.000040834664105204865, -0.1532757431268692, 0.03312944993376732, 0.037237878888845444, 0.006582805421203375, -0.018114112317562103, -0.04424576088786125, -0.10179384052753448, 0.12925361096858978, -0.05290345475077629, -0.053517285734415054, -0.07407750189304352, -0.056111112236976624, 0.11954375356435776, -0.051811717450618744, 0.05964844673871994, -0.03253856301307678, -0.03517496585845947, -0.08935095369815826, -0.09134237468242645, 0.12569497525691986, -0.15165700018405914, 0.008372670970857143, -0.10362232476472855, 0.07550536096096039, -0.06283175945281982, 0.040086161345243454, 0.01594039797782898, 0.015317576937377453, -0.05658023804426193, -0.04569414258003235, 0.0241472776979208, -0.07941046357154846, 0.1317061334848404, -0.012372943572700024, -0.09904076904058456, -0.06987979263067245, 0.018766043707728386, -0.08652849495410919, 0.13333114981651306, 0.2562698423862457, -0.05168404057621956, 0.11424782872200012, 0.2733060419559479, -0.1169561818242073, -0.22521817684173584, -0.1369016319513321, -0.12080144137144089, -0.018031004816293716, -0.057368241250514984, -0.16948819160461426, 0.03436417877674103, 0.15529362857341766, -0.08302679657936096, 0.0850416049361229, -0.24666070938110352, -0.05932839587330818, 0.1503632664680481, -0.0030005150474607944, 0.41653740406036377, -0.11908061057329178, -0.09709174185991287, -0.10717098414897919, -0.1444958746433258, 0.1094001904129982, -0.05136117339134216, 0.037704531103372574, -0.03598437085747719, -0.04222622513771057, -0.009638707153499126, -0.021427446976304054, 0.1637425273656845, -0.08329187333583832, -0.010542508214712143, -0.06988102942705154, -0.03455633297562599, 0.12281136214733124, -0.007728678174316883, -0.017965659499168396, -0.06059487536549568, -0.004578818567097187, -0.13814301788806915, -0.0037811617366969585, -0.07760240882635117, 0.08373400568962097, -0.019743552431464195, -0.02651406079530716, -0.023969998583197594, -0.0009343305719085038, 0.0057798526249825954, -0.006028153467923403, 0.16817377507686615, 0.033949077129364014, 0.09371142834424973, 0.14244361221790314, 0.04760768637061119, -0.09897590428590775, 0.010091078467667103, -0.040613092482089996, -0.06316415965557098, 0.07718313485383987, -0.036711275577545166, 0.011431853286921978, 0.12247196584939957, -0.04685593023896217, 0.024640360847115517, 0.04018067196011543, -0.033217839896678925, -0.0193686094135046, 0.15400540828704834, -0.10077078640460968, 0.07469917833805084, 0.014176069758832455, 0.04720723628997803, 0.08933867514133453, -0.017212139442563057, 0.10418465733528137, 0.04096699506044388, -0.0545668825507164, 0.01672716811299324, -0.024263354018330574, -0.04546774551272392, 0.09527066349983215, 0.06885433942079544, -0.00006753960042260587, -0.10021428763866425, 0.09361714124679565, -0.00035994272911921144, -0.18267902731895447, 0.005553741008043289, 0.09958399832248688, -0.05338464304804802, -0.09836888313293457, 0.044450175017118454, 0.09538974612951279, -0.2777233123779297, -0.12030315399169922, -0.14167824387550354, -0.07271117717027664, 0.06314882636070251, 0.21333011984825134, 0.05926893651485443, 0.015700865536928177, 0.021206650882959366, -0.027455834671854973, 0.04960489273071289, -0.009198910556733608, -0.05697852745652199, 0.024983102455735207, -0.04658493399620056, -0.07104982435703278, -0.003042182419449091, 0.028430908918380737, -0.04319439083337784, 0.04843815043568611, -0.1761520653963089, 0.046321313828229904, -0.13682974874973297, 0.04149448499083519, -0.08104518055915833, -0.044697873294353485, -0.0419221892952919, -0.08323553204536438, -0.03899591043591499, -0.06855668127536774, -0.0966012254357338, 0.004618645645678043, 0.02510862797498703, 0.0683593899011612, -0.08092878013849258, -0.054927758872509, 0.07724858075380325, -0.011857801117002964, 0.05876767635345459, 0.07942426204681396, -0.002903212094679475, 0.10072943568229675, -0.1843300312757492, -0.04844437912106514, 0.07801813632249832, 0.049724314361810684, 0.05745372548699379, -0.0032924681436270475, 0.011291739530861378, 0.10495751351118088, -0.015450369566679, 0.0484817810356617, -0.02553027868270874, -0.1281830072402954, 0.0280994214117527, 0.056890591979026794, -0.1978885978460312, 0.03488048538565636, -0.021329576149582863, 0.1822541058063507, -0.05559267848730087, 0.07007995992898941, -0.060960303992033005, -0.0026777959428727627, -0.03146453946828842, 0.028087852522730827, -0.01897263340651989, -0.09729664027690887, -0.08406274765729904, -0.02144882082939148, -0.022760894149541855, 0.012463339604437351, 0.24773356318473816, 0.02878081612288952, -0.022428365424275398, 0.027011040598154068, 0.015260271728038788, -0.010643012821674347, 0.013154328800737858, 0.23743028938770294, 0.06911129504442215, -0.01155712641775608, -0.139927476644516, 0.08704622834920883, 0.03095470555126667, 0.0397803857922554, 0.026455407962203026, 0.1462932825088501, 0.2702261805534363, 0.07365474849939346, 0.05616656690835953, -0.028700444847345352, 0.0015965357888489962, -0.08856857568025589, 0.0730258971452713, 0.030638987198472023, 0.07038411498069763, 0.07882748544216156, 0.16647955775260925, -0.04575353488326073, 0.038287799805402756, -0.030874252319335938, 0.022410571575164795, -0.11511614173650742, -0.09438202530145645, -0.03993077576160431, -0.0910993367433548, -0.01092811580747366, -0.06014535203576088, 0.04263549670577049, 0.024602459743618965, 0.023475442081689835, -0.06346417963504791, -0.07940821349620819, 0.06104256212711334, -0.08281167596578598, 0.01737947016954422, 0.03951617330312729, 0.09022513031959534, -0.09439235180616379, 0.022808320820331573, -0.0862647294998169, 0.058885328471660614, -0.0434952974319458, 0.05588072910904884, 0.003622537013143301, 0.009331348352134228, -0.05551844462752342, -0.03616322949528694, -0.030741186812520027, 0.05308174341917038, 0.017511676996946335, 0.1843118518590927, -0.0061880419962108135, 0.039190638810396194, 0.06121756508946419, 0.20344707369804382, -0.022450828924775124, -0.06977364420890808, -0.04378997161984444, 0.07879894971847534, 0.015052388422191143, 0.0980411171913147, -0.013988901861011982, 0.0012219490017741919, -0.05717403069138527, 0.2495289444923401, 0.2928844392299652, -0.0397014245390892, -0.008310964331030846, 0.04620962589979172, 0.02411886677145958, 0.08754832297563553, 0.05432867258787155, 0.09651612490415573, 0.2969628572463989, -0.0976187065243721, -0.07883062958717346, -0.09446041285991669, 0.07065851241350174, -0.12698574364185333, 0.09976290911436081, -0.042604461312294006, -0.09925293177366257, -0.009981575421988964, 0.10233666747808456, -0.09637825191020966, -0.04688365012407303, 0.0474868044257164, -0.19230656325817108, -0.09537660330533981, -0.022195814177393913, 0.13683685660362244, 0.05986500903964043, 0.0628414899110794, -0.02302134968340397, -0.05936474725604057, 0.0068147312849760056, 0.0038668832276016474, -0.1779661327600479, -0.15910528600215912, 0.04968655854463577, 0.01929519511759281, 0.2247513234615326, -0.025113116949796677, 0.025674039497971535, 0.11185403913259506, 0.05785488337278366, -0.10148259997367859, 0.03135709837079048, 0.06966892629861832, -0.034127138555049896, 0.05687996372580528, -0.1347975879907608, 0.010469846427440643, -0.031402587890625, 0.05596665292978287, -0.053629446774721146, 0.05869147926568985, 0.10633812099695206, -0.04445892944931984, -0.049189966171979904, 0.11179493367671967, -0.10284868627786636, 0.02670038677752018, 0.0482979491353035, -0.0069682784378528595, -0.04284790903329849, -0.032642826437950134, -0.008706665597856045, 0.08358263969421387, -0.03475720435380936, -0.0770348310470581, -0.07667385786771774, -0.02352266199886799, 0.05469944700598717, 0.000598185695707798, -0.15249764919281006, 0.019031433388590813, -0.10293423384428024, -0.00837673433125019, -0.1356157511472702, -0.011688479222357273, 0.006212752312421799, 0.018764175474643707, 0.012203912250697613, -0.0772874727845192, -0.00635465607047081, 0.005130188073962927, -0.04268669709563255, -0.06864291429519653 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/roberta-base-ca # BERTa: RoBERTa-based Catalan language model ## BibTeX citation If you use any of these resources (datasets or models) in your work, please cite our latest paper: ```bibtex @inproceedings{armengol-estape-etal-2021-multilingual, title = "Are Multilingual Models the Best Choice for Moderately Under-resourced Languages? {A} Comprehensive Assessment for {C}atalan", author = "Armengol-Estap{\'e}, Jordi and Carrino, Casimiro Pio and Rodriguez-Penagos, Carlos and de Gibert Bonet, Ona and Armentano-Oller, Carme and Gonzalez-Agirre, Aitor and Melero, Maite and Villegas, Marta", booktitle = "Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021", month = aug, year = "2021", address = "Online", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2021.findings-acl.437", doi = "10.18653/v1/2021.findings-acl.437", pages = "4933--4946", } ``` ## Model description BERTa is a transformer-based masked language model for the Catalan language. It is based on the [RoBERTA](https://github.com/pytorch/fairseq/tree/master/examples/roberta) base model and has been trained on a medium-size corpus collected from publicly available corpora and crawlers. ## Training corpora and preprocessing The training corpus consists of several corpora gathered from web crawling and public corpora. The publicly available corpora are: 1. the Catalan part of the [DOGC](http://opus.nlpl.eu/DOGC-v2.php) corpus, a set of documents from the Official Gazette of the Catalan Government 2. the [Catalan Open Subtitles](http://opus.nlpl.eu/download.php?f=OpenSubtitles/v2018/mono/OpenSubtitles.raw.ca.gz), a collection of translated movie subtitles 3. the non-shuffled version of the Catalan part of the [OSCAR](https://traces1.inria.fr/oscar/) corpus \\\\cite{suarez2019asynchronous}, a collection of monolingual corpora, filtered from [Common Crawl](https://commoncrawl.org/about/) 4. The [CaWac](http://nlp.ffzg.hr/resources/corpora/cawac/) corpus, a web corpus of Catalan built from the .cat top-level-domain in late 2013 the non-deduplicated version 5. the [Catalan Wikipedia articles](https://ftp.acc.umu.se/mirror/wikimedia.org/dumps/cawiki/20200801/) downloaded on 18-08-2020. The crawled corpora are: 6. The Catalan General Crawling, obtained by crawling the 500 most popular .cat and .ad domains 7. the Catalan Government Crawling, obtained by crawling the .gencat domain and subdomains, belonging to the Catalan Government 8. the ACN corpus with 220k news items from March 2015 until October 2020, crawled from the [Catalan News Agency](https://www.acn.cat/) To obtain a high-quality training corpus, each corpus have preprocessed with a pipeline of operations, including among the others, sentence splitting, language detection, filtering of bad-formed sentences and deduplication of repetitive contents. During the process, we keep document boundaries are kept. Finally, the corpora are concatenated and further global deduplication among the corpora is applied. The final training corpus consists of about 1,8B tokens. ## Tokenization and pretraining The training corpus has been tokenized using a byte version of [Byte-Pair Encoding (BPE)](https://github.com/openai/gpt-2) used in the original [RoBERTA](https://github.com/pytorch/fairseq/tree/master/examples/roberta) model with a vocabulary size of 52,000 tokens. The BERTa pretraining consists of a masked language model training that follows the approach employed for the RoBERTa base model with the same hyperparameters as in the original work. The training lasted a total of 48 hours with 16 NVIDIA V100 GPUs of 16GB DDRAM. ## Evaluation ## CLUB benchmark The BERTa model has been fine-tuned on the downstream tasks of the Catalan Language Understanding Evaluation benchmark (CLUB), that has been created along with the model. It contains the following tasks and their related datasets: 1. Part-of-Speech Tagging (POS) Catalan-Ancora: from the [Universal Dependencies treebank](https://github.com/UniversalDependencies/UD_Catalan-AnCora) of the well-known Ancora corpus 2. Named Entity Recognition (NER) **[AnCora Catalan 2.0.0](https://zenodo.org/record/4762031#.YKaFjqGxWUk)**: extracted named entities from the original [Ancora](https://doi.org/10.5281/zenodo.4762030) version, filtering out some unconventional ones, like book titles, and transcribed them into a standard CONLL-IOB format 3. Text Classification (TC) **[TeCla](https://doi.org/10.5281/zenodo.4627197)**: consisting of 137k news pieces from the Catalan News Agency ([ACN](https://www.acn.cat/)) corpus 4. Semantic Textual Similarity (STS) **[Catalan semantic textual similarity](https://doi.org/10.5281/zenodo.4529183)**: consisting of more than 3000 sentence pairs, annotated with the semantic similarity between them, scraped from the [Catalan Textual Corpus](https://doi.org/10.5281/zenodo.4519349) 5. Question Answering (QA): **[ViquiQuAD](https://doi.org/10.5281/zenodo.4562344)**: consisting of more than 15,000 questions outsourced from Catalan Wikipedia randomly chosen from a set of 596 articles that were originally written in Catalan. **[XQuAD](https://doi.org/10.5281/zenodo.4526223)**: the Catalan translation of XQuAD, a multilingual collection of manual translations of 1,190 question-answer pairs from English Wikipedia used only as a _test set_ Here are the train/dev/test splits of the datasets: | Task (Dataset) | Total | Train | Dev | Test | |:--|:--|:--|:--|:--| | NER (Ancora) |13,581 | 10,628 | 1,427 | 1,526 | | POS (Ancora)| 16,678 | 13,123 | 1,709 | 1,846 | | STS | 3,073 | 2,073 | 500 | 500 | | TC (TeCla) | 137,775 | 110,203 | 13,786 | 13,786| | QA (ViquiQuAD) | 14,239 | 11,255 | 1,492 | 1,429 | _The fine-tuning on downstream tasks have been performed with the HuggingFace [**Transformers**](https://github.com/huggingface/transformers) library_ ## Results Below the evaluation results on the CLUB tasks compared with the multilingual mBERT, XLM-RoBERTa models and the Catalan WikiBERT-ca model | Task | NER (F1) | POS (F1) | STS (Pearson) | TC (accuracy) | QA (ViquiQuAD) (F1/EM) | QA (XQuAD) (F1/EM) | | ------------|:-------------:| -----:|:------|:-------|:------|:----| | BERTa | **88.13** | **98.97** | **79.73** | **74.16** | **86.97/72.29** | **68.89/48.87** | | mBERT | 86.38 | 98.82 | 76.34 | 70.56 | 86.97/72.22 | 67.15/46.51 | | XLM-RoBERTa | 87.66 | 98.89 | 75.40 | 71.68 | 85.50/70.47 | 67.10/46.42 | | WikiBERT-ca | 77.66 | 97.60 | 77.18 | 73.22 | 85.45/70.75 | 65.21/36.60 | ## Intended uses & limitations The model is ready-to-use only for masked language modelling to perform the Fill Mask task (try the inference API or read the next section) However, the is intended to be fine-tuned on non-generative downstream tasks such as Question Answering, Text Classification or Named Entity Recognition. --- ## Using BERTa ## Load model and tokenizer ``` python from transformers import AutoTokenizer, AutoModelForMaskedLM tokenizer = AutoTokenizer.from_pretrained("BSC-TeMU/roberta-base-ca-cased") model = AutoModelForMaskedLM.from_pretrained("BSC-TeMU/roberta-base-ca-cased") ``` ## Fill Mask task Below, an example of how to use the masked language modelling task with a pipeline. ```python >>> from transformers import pipeline >>> unmasker = pipeline('fill-mask', model='BSC-TeMU/roberta-base-ca-cased') >>> unmasker("Situada a la costa de la mar Mediterrània, <mask> s'assenta en una plana formada " "entre els deltes de les desembocadures dels rius Llobregat, al sud-oest, " "i Besòs, al nord-est, i limitada pel sud-est per la línia de costa," "i pel nord-oest per la serralada de Collserola " "(amb el cim del Tibidabo, 516,2 m, com a punt més alt) que segueix paral·lela " "la línia de costa encaixant la ciutat en un perímetre molt definit.") [ { "sequence": " Situada a la costa de la mar Mediterrània, <mask> s'assenta en una plana formada " "entre els deltes de les desembocadures dels rius Llobregat, al sud-oest, " "i Besòs, al nord-est, i limitada pel sud-est per la línia de costa," "i pel nord-oest per la serralada de Collserola " "(amb el cim del Tibidabo, 516,2 m, com a punt més alt) que segueix paral·lela " "la línia de costa encaixant la ciutat en un perímetre molt definit.", "score": 0.4177263379096985, "token": 734, "token_str": " Barcelona" }, { "sequence": " Situada a la costa de la mar Mediterrània, <mask> s'assenta en una plana formada " "entre els deltes de les desembocadures dels rius Llobregat, al sud-oest, " "i Besòs, al nord-est, i limitada pel sud-est per la línia de costa," "i pel nord-oest per la serralada de Collserola " "(amb el cim del Tibidabo, 516,2 m, com a punt més alt) que segueix paral·lela " "la línia de costa encaixant la ciutat en un perímetre molt definit.", "score": 0.10696165263652802, "token": 3849, "token_str": " Badalona" }, { "sequence": " Situada a la costa de la mar Mediterrània, <mask> s'assenta en una plana formada " "entre els deltes de les desembocadures dels rius Llobregat, al sud-oest, " "i Besòs, al nord-est, i limitada pel sud-est per la línia de costa," "i pel nord-oest per la serralada de Collserola " "(amb el cim del Tibidabo, 516,2 m, com a punt més alt) que segueix paral·lela " "la línia de costa encaixant la ciutat en un perímetre molt definit.", "score": 0.08135009557008743, "token": 19349, "token_str": " Collserola" }, { "sequence": " Situada a la costa de la mar Mediterrània, <mask> s'assenta en una plana formada " "entre els deltes de les desembocadures dels rius Llobregat, al sud-oest, " "i Besòs, al nord-est, i limitada pel sud-est per la línia de costa," "i pel nord-oest per la serralada de Collserola " "(amb el cim del Tibidabo, 516,2 m, com a punt més alt) que segueix paral·lela " "la línia de costa encaixant la ciutat en un perímetre molt definit.", "score": 0.07330769300460815, "token": 4974, "token_str": " Terrassa" }, { "sequence": " Situada a la costa de la mar Mediterrània, <mask> s'assenta en una plana formada " "entre els deltes de les desembocadures dels rius Llobregat, al sud-oest, " "i Besòs, al nord-est, i limitada pel sud-est per la línia de costa," "i pel nord-oest per la serralada de Collserola " "(amb el cim del Tibidabo, 516,2 m, com a punt més alt) que segueix paral·lela " "la línia de costa encaixant la ciutat en un perímetre molt definit.", "score": 0.03317456692457199, "token": 14333, "token_str": " Gavà" } ] ``` This model was originally published as [bsc/roberta-base-ca-cased](https://huggingface.co/bsc/roberta-base-ca-cased).
{"language": "ca", "license": "apache-2.0", "tags": ["masked-lm", "BERTa", "catalan"], "widget": [{"text": "El Catal\u00e0 \u00e9s una llengua molt <mask>."}, {"text": "Salvador Dal\u00ed va viure a <mask>."}, {"text": "La Costa Brava t\u00e9 les millors <mask> d'Espanya."}, {"text": "El cacaolat \u00e9s un batut de <mask>."}, {"text": "<mask> \u00e9s la capital de la Garrotxa."}, {"text": "Vaig al <mask> a buscar bolets."}, {"text": "Antoni Gaud\u00ed vas ser un <mask> molt important per la ciutat."}, {"text": "Catalunya \u00e9s una refer\u00e8ncia en <mask> a nivell europeu."}]}
fill-mask
BSC-LT/roberta-base-ca
[ "transformers", "pytorch", "roberta", "fill-mask", "masked-lm", "BERTa", "catalan", "ca", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[ "ca" ]
TAGS #transformers #pytorch #roberta #fill-mask #masked-lm #BERTa #catalan #ca #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL BERTa: RoBERTa-based Catalan language model =========================================== BibTeX citation --------------- If you use any of these resources (datasets or models) in your work, please cite our latest paper: Model description ----------------- BERTa is a transformer-based masked language model for the Catalan language. It is based on the RoBERTA base model and has been trained on a medium-size corpus collected from publicly available corpora and crawlers. Training corpora and preprocessing ---------------------------------- The training corpus consists of several corpora gathered from web crawling and public corpora. The publicly available corpora are: 1. the Catalan part of the DOGC corpus, a set of documents from the Official Gazette of the Catalan Government 2. the Catalan Open Subtitles, a collection of translated movie subtitles 3. the non-shuffled version of the Catalan part of the OSCAR corpus \\cite{suarez2019asynchronous}, a collection of monolingual corpora, filtered from Common Crawl 4. The CaWac corpus, a web corpus of Catalan built from the .cat top-level-domain in late 2013 the non-deduplicated version 5. the Catalan Wikipedia articles downloaded on 18-08-2020. The crawled corpora are: 6. The Catalan General Crawling, obtained by crawling the 500 most popular .cat and .ad domains 7. the Catalan Government Crawling, obtained by crawling the .gencat domain and subdomains, belonging to the Catalan Government 8. the ACN corpus with 220k news items from March 2015 until October 2020, crawled from the Catalan News Agency To obtain a high-quality training corpus, each corpus have preprocessed with a pipeline of operations, including among the others, sentence splitting, language detection, filtering of bad-formed sentences and deduplication of repetitive contents. During the process, we keep document boundaries are kept. Finally, the corpora are concatenated and further global deduplication among the corpora is applied. The final training corpus consists of about 1,8B tokens. Tokenization and pretraining ---------------------------- The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original RoBERTA model with a vocabulary size of 52,000 tokens. The BERTa pretraining consists of a masked language model training that follows the approach employed for the RoBERTa base model with the same hyperparameters as in the original work. The training lasted a total of 48 hours with 16 NVIDIA V100 GPUs of 16GB DDRAM. Evaluation ---------- CLUB benchmark -------------- The BERTa model has been fine-tuned on the downstream tasks of the Catalan Language Understanding Evaluation benchmark (CLUB), that has been created along with the model. It contains the following tasks and their related datasets: 1. Part-of-Speech Tagging (POS) Catalan-Ancora: from the Universal Dependencies treebank of the well-known Ancora corpus 2. Named Entity Recognition (NER) AnCora Catalan 2.0.0: extracted named entities from the original Ancora version, filtering out some unconventional ones, like book titles, and transcribed them into a standard CONLL-IOB format 3. Text Classification (TC) TeCla: consisting of 137k news pieces from the Catalan News Agency (ACN) corpus 4. Semantic Textual Similarity (STS) Catalan semantic textual similarity: consisting of more than 3000 sentence pairs, annotated with the semantic similarity between them, scraped from the Catalan Textual Corpus 5. Question Answering (QA): ViquiQuAD: consisting of more than 15,000 questions outsourced from Catalan Wikipedia randomly chosen from a set of 596 articles that were originally written in Catalan. XQuAD: the Catalan translation of XQuAD, a multilingual collection of manual translations of 1,190 question-answer pairs from English Wikipedia used only as a *test set* Here are the train/dev/test splits of the datasets: *The fine-tuning on downstream tasks have been performed with the HuggingFace Transformers library* Results ------- Below the evaluation results on the CLUB tasks compared with the multilingual mBERT, XLM-RoBERTa models and the Catalan WikiBERT-ca model Intended uses & limitations --------------------------- The model is ready-to-use only for masked language modelling to perform the Fill Mask task (try the inference API or read the next section) However, the is intended to be fine-tuned on non-generative downstream tasks such as Question Answering, Text Classification or Named Entity Recognition. --- Using BERTa ----------- Load model and tokenizer ------------------------ Fill Mask task -------------- Below, an example of how to use the masked language modelling task with a pipeline. This model was originally published as bsc/roberta-base-ca-cased.
[]
[ "TAGS\n#transformers #pytorch #roberta #fill-mask #masked-lm #BERTa #catalan #ca #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 58 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #fill-mask #masked-lm #BERTa #catalan #ca #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.09753497689962387, 0.06666616350412369, -0.0074958186596632, 0.07030793279409409, 0.06641024351119995, -0.03811996802687645, 0.1164112240076065, 0.0803503692150116, 0.02626921609044075, 0.014110668562352657, 0.1832338273525238, 0.2462538778781891, -0.01086397934705019, 0.01738467998802662, -0.035390306264162064, -0.20465223491191864, 0.09526652097702026, 0.001188041758723557, -0.09380116313695908, 0.08792036026716232, 0.09558574855327606, -0.002196772489696741, 0.07978225499391556, 0.008836928755044937, -0.027860401198267937, 0.05043594911694527, 0.037189383059740067, -0.12265179306268692, 0.12749072909355164, 0.06368418037891388, 0.11280244588851929, 0.023251038044691086, -0.03839481994509697, -0.10388507694005966, 0.02887277863919735, -0.020138876512646675, -0.0855686292052269, 0.053684789687395096, 0.023237895220518112, -0.08621010184288025, 0.015185720287263393, 0.02935798093676567, 0.004266793839633465, 0.03141910582780838, -0.1269156038761139, -0.21249595284461975, -0.05048419162631035, 0.020250609144568443, 0.07222558557987213, 0.06579746305942535, 0.03659190610051155, 0.11670172959566116, -0.1296110302209854, 0.0353250578045845, 0.15748727321624756, -0.31808626651763916, -0.025643788278102875, -0.050515688955783844, 0.07042812556028366, -0.03352222591638565, 0.03862294927239418, 0.07698079943656921, 0.07040172070264816, -0.0019792327657341957, -0.013971477746963501, -0.07524652034044266, -0.1210571825504303, -0.01020940113812685, -0.040484409779310226, -0.08408394455909729, 0.16045045852661133, -0.0339314229786396, 0.020976878702640533, 0.015794266015291214, -0.0903872549533844, 0.03612701967358589, -0.033228255808353424, 0.00813906267285347, 0.036270562559366226, 0.07662171870470047, 0.07090990245342255, 0.021273622289299965, -0.10745826363563538, 0.02918236516416073, -0.24915383756160736, 0.1479649692773819, 0.043743062764406204, 0.06132875382900238, -0.10306750237941742, 0.03497820720076561, -0.020661795511841774, -0.1076556146144867, -0.009386702440679073, -0.002274835016578436, 0.09294047951698303, 0.056383054703474045, -0.05311908572912216, 0.015708882361650467, 0.12746775150299072, 0.2760430574417114, -0.010497733019292355, -0.0017016284400597215, -0.016204502433538437, 0.14265359938144684, -0.03605453297495842, 0.04445668309926987, -0.010885747149586678, -0.013922439888119698, 0.061269793659448624, -0.15770122408866882, 0.09645963460206985, -0.027002405375242233, -0.19935786724090576, -0.06750202178955078, -0.058763232082128525, 0.13698726892471313, 0.08333276212215424, 0.01905413344502449, -0.10605717450380325, 0.019770639017224312, 0.1395062804222107, -0.04736700281500816, 0.018946604803204536, -0.0321563296020031, 0.03960906341671944, -0.002538809785619378, 0.020639991387724876, 0.007638399954885244, -0.01351094152778387, 0.07184555381536484, -0.06293779611587524, -0.05172457918524742, -0.021592089906334877, -0.016897747293114662, 0.10084279626607895, -0.14041270315647125, 0.05047712102532387, -0.17659832537174225, -0.10942380875349045, 0.05781269446015358, 0.0952693521976471, -0.0546678863465786, -0.019227100536227226, -0.022395560517907143, 0.024759715422987938, 0.021434329450130463, -0.07640956342220306, -0.04487571120262146, -0.07022382318973541, 0.0837714821100235, -0.03150877729058266, 0.079439178109169, -0.19327765703201294, 0.029928358271718025, -0.10090310126543045, 0.051123425364494324, -0.1177314817905426, -0.030952230095863342, -0.03380122035741806, 0.13167859613895416, -0.06435725837945938, -0.03318313509225845, -0.026402302086353302, 0.06618278473615646, 0.03067123144865036, 0.1370547115802765, -0.08334065973758698, -0.0954110324382782, 0.2327098697423935, -0.08980602771043777, -0.14849582314491272, 0.10516543686389923, 0.028545644134283066, 0.00994723942130804, 0.04326673969626427, 0.07837753742933273, -0.015358515083789825, -0.12709100544452667, 0.09280353784561157, 0.07165738195180893, -0.10730606317520142, -0.18657554686069489, 0.11447582393884659, -0.08392664045095444, -0.09142836928367615, 0.049755144864320755, -0.033636342734098434, 0.07947494089603424, -0.00816990528255701, -0.09419496357440948, -0.008559523150324821, -0.04573169723153114, 0.05096253380179405, 0.01142403855919838, 0.06468845903873444, -0.08782726526260376, -0.001578991417773068, -0.08527370542287827, 0.000370480353012681, 0.08037739992141724, 0.06155849248170853, -0.10680156946182251, 0.16226190328598022, -0.05290649086236954, -0.0019871757831424475, -0.10648804903030396, 0.02266785316169262, -0.05636056885123253, 0.016141526401042938, 0.0008058485691435635, 0.05318402871489525, 0.07009688764810562, -0.04397676885128021, -0.029405465349555016, 0.027921592816710472, 0.10148077458143234, 0.011707134544849396, 0.040038224309682846, -0.17335687577724457, 0.06386362761259079, -0.03485743701457977, 0.017756959423422813, 0.027339739724993706, 0.0032631519716233015, 0.039063915610313416, 0.05935205891728401, -0.06321962177753448, 0.07739057391881943, -0.10425035655498505, 0.04350198432803154, -0.07013088464736938, 0.016908397898077965, 0.14091037213802338, 0.01477771531790495, -0.07136497646570206, 0.18923614919185638, -0.0618162527680397, 0.2803896367549896, 0.17967896163463593, -0.21159009635448456, 0.0014976481907069683, 0.04578915983438492, -0.0305510051548481, 0.00430348701775074, 0.04906390607357025, -0.002241667127236724, 0.03780189901590347, -0.01576586253941059, 0.15491098165512085, -0.10303021967411041, -0.006806542165577412, 0.04335000738501549, -0.08067258447408676, -0.07585309445858002, 0.09071891754865646, 0.169534832239151, -0.05659138038754463, 0.18362976610660553, 0.3263653814792633, -0.008000096306204796, 0.15238535404205322, -0.013800548389554024, 0.014924257062375546, -0.020495738834142685, 0.014270511455833912, 0.0316571369767189, 0.08381951600313187, -0.16788403689861298, -0.004877978004515171, 0.04354294016957283, -0.029917113482952118, 0.03268497809767723, -0.13588006794452667, -0.07036209106445312, 0.018152346834540367, -0.0019113152520731091, -0.08125686645507812, 0.10351737588644028, -0.040628716349601746, 0.08243098109960556, 0.0360378660261631, -0.19458234310150146, 0.13457080721855164, 0.004231925588101149, -0.042521122843027115, 0.13951300084590912, -0.17855283617973328, -0.2952556610107422, -0.14027109742164612, -0.13644596934318542, -0.015241514891386032, 0.05864879488945007, 0.08738280832767487, -0.07239104807376862, -0.022382183000445366, 0.08412977308034897, -0.02072072960436344, 0.012533731758594513, 0.009382450953125954, -0.058729108422994614, 0.041373323649168015, -0.021714407950639725, -0.09835600107908249, -0.05614583194255829, 0.01290005911141634, -0.07814329862594604, 0.034456513822078705, -0.09065163880586624, 0.1189415231347084, 0.08831143379211426, 0.03497492894530296, 0.029079198837280273, -0.008308750577270985, 0.15289947390556335, -0.08562420308589935, -0.012319410219788551, 0.17064934968948364, 0.023038683459162712, 0.05322656035423279, 0.21305561065673828, 0.05175269767642021, -0.06772402673959732, -0.04461659863591194, -0.06337771564722061, -0.1208646297454834, -0.19617734849452972, -0.12846896052360535, -0.09260060638189316, 0.018008818849921227, 0.0615839958190918, 0.06990565359592438, 0.15211470425128937, 0.1053951233625412, 0.034352414309978485, -0.0418853759765625, -0.060492031276226044, 0.053230755031108856, 0.20422303676605225, -0.03112088330090046, 0.08986455202102661, -0.07502397149801254, -0.1114310622215271, 0.11523742228746414, 0.1060168519616127, 0.08343582600355148, 0.13569095730781555, 0.01000299584120512, 0.11050348728895187, 0.18785546720027924, 0.038240768015384674, 0.08774429559707642, 0.033135682344436646, -0.03736535832285881, -0.059859439730644226, -0.034897007048130035, -0.045509908348321915, -0.01143561489880085, 0.03895621374249458, -0.10130122303962708, -0.05036628246307373, -0.21544167399406433, 0.05996793881058693, 0.18899166584014893, 0.04364914447069168, -0.15022341907024384, 0.039530687034130096, 0.07240013778209686, 0.01576373353600502, -0.048780910670757294, 0.06624454259872437, -0.03697482869029045, -0.14555691182613373, 0.12080507725477219, -0.019298093393445015, 0.08621416985988617, 0.04098563641309738, 0.07579993456602097, -0.05234168469905853, -0.07318101078271866, 0.07581759989261627, 0.07637543231248856, -0.2693822383880615, 0.3309488296508789, 0.0180331002920866, -0.021762410178780556, -0.08117172122001648, -0.00826634094119072, 0.03078647516667843, 0.18291181325912476, 0.17347268760204315, 0.052213914692401886, -0.0787481963634491, -0.04416564106941223, -0.02573743276298046, 0.04770680516958237, -0.017900962382555008, -0.04062040522694588, -0.06978767365217209, -0.02510971948504448, -0.052548471838235855, -0.01603890396654606, 0.07524718344211578, 0.006449087988585234, -0.15427863597869873, 0.05474703013896942, 0.120358407497406, -0.0023204495664685965, -0.04361971840262413, -0.04368216171860695, -0.13703693449497223, 0.160131573677063, -0.10827888548374176, -0.039956070482730865, -0.06112771853804588, -0.18010593950748444, 0.045444563031196594, -0.0673278197646141, 0.08927146345376968, -0.0679147019982338, -0.031276289373636246, -0.09961669892072678, -0.09483065456151962, 0.16783462464809418, -0.1503002792596817, -0.012070149183273315, -0.07250642031431198, 0.10235901921987534, -0.04725952446460724, 0.007782539818435907, 0.029204772785305977, 0.027469731867313385, -0.06756112724542618, -0.07524224370718002, -0.02265756018459797, -0.08040329813957214, 0.022438431158661842, 0.010282250121235847, -0.08372225612401962, -0.07046295702457428, 0.05151043459773064, -0.05713193863630295, 0.1569886952638626, 0.2705211341381073, -0.0625411868095398, 0.16665087640285492, 0.2645893096923828, -0.07574088871479034, -0.3222341239452362, -0.1717805117368698, -0.1661396622657776, -0.036928433924913406, 0.0017275098944082856, -0.15999078750610352, 0.0280136875808239, 0.08692464977502823, -0.09518902748823166, 0.10664052516222, -0.27781376242637634, -0.09833042323589325, 0.20809772610664368, 0.00736171705648303, 0.47870051860809326, -0.06464923173189163, -0.08638697117567062, -0.10780258476734161, -0.09196368604898453, 0.05788812413811684, -0.025658709928393364, 0.0792386457324028, -0.022669076919555664, -0.022683732211589813, -0.0012941042659804225, -0.0390610471367836, 0.1516573131084442, -0.02027764730155468, 0.010350039228796959, -0.09023687988519669, -0.012416070327162743, 0.14010630548000336, 0.0077315568923950195, -0.01730651594698429, -0.08051510900259018, -0.038920026272535324, -0.031082967296242714, -0.0009715181076899171, -0.10229107737541199, 0.10015779733657837, -0.021852919831871986, -0.01753867045044899, -0.013278217054903507, 0.0016432591946795583, 0.03901790827512741, 0.011358136311173439, 0.1600855141878128, 0.000923116342164576, 0.08778399974107742, 0.0695982277393341, 0.04839107766747475, -0.15453855693340302, -0.002930206246674061, -0.0498744435608387, -0.06496979296207428, 0.07174867391586304, -0.01654074713587761, 0.023537028580904007, 0.08374663442373276, -0.06528123468160629, 0.015709122642874718, 0.06519676744937897, -0.01577683910727501, -0.01566530391573906, 0.16952835023403168, -0.10373403877019882, -0.028128081932663918, 0.01598433032631874, 0.02221347764134407, 0.11452396214008331, 0.017211979255080223, 0.06873276084661484, 0.06297538429498672, -0.02999212220311165, 0.018080441281199455, 0.01039337832480669, -0.10448592156171799, 0.03820427134633064, 0.03541351482272148, -0.014274570159614086, -0.12617428600788116, 0.039890747517347336, -0.020320912823081017, -0.18203015625476837, 0.007781743537634611, 0.0783788412809372, -0.0885029137134552, -0.11205068975687027, 0.027403661981225014, 0.06671711057424545, -0.2213466614484787, -0.15027455985546112, -0.10773720592260361, -0.1342393010854721, 0.06915983557701111, 0.18182092905044556, 0.09409870207309723, 0.060937706381082535, 0.03124557062983513, -0.06633257865905762, 0.053611207753419876, -0.010592061094939709, -0.04456663876771927, 0.006540765520185232, -0.018403103575110435, -0.07772190123796463, 0.013621602207422256, 0.11756031215190887, -0.06483542174100876, -0.011561108753085136, -0.13695897161960602, 0.057773418724536896, -0.13559503853321075, -0.054521575570106506, -0.06857680529356003, -0.05082455649971962, 0.010756142437458038, -0.07715780287981033, -0.02427823282778263, -0.03687863424420357, -0.11383743584156036, 0.014379681088030338, 0.03832392767071724, 0.030175456777215004, -0.10369787365198135, -0.04563833028078079, 0.12285907566547394, -0.020473558455705643, 0.06702954322099686, 0.050256602466106415, -0.05181357264518738, 0.08030492812395096, -0.2107379287481308, -0.07760443538427353, 0.051520269364118576, 0.028476694598793983, 0.05301763117313385, 0.009143684059381485, 0.01864769496023655, 0.12127711623907089, 0.011042500846087933, 0.06999951601028442, 0.054109495133161545, -0.12419168651103973, 0.04298299923539162, 0.07636667042970657, -0.20202605426311493, 0.029686886817216873, -0.04374835640192032, 0.11992371827363968, -0.04617039114236832, 0.1064206063747406, -0.06309348344802856, 0.018745454028248787, 0.0323416031897068, 0.020698435604572296, -0.023877399042248726, -0.11835213750600815, -0.07948987185955048, -0.04104692116379738, -0.05860573798418045, 0.006513563916087151, 0.23517383635044098, 0.024219047278165817, 0.01213857065886259, 0.05369839072227478, 0.10271546244621277, -0.08374990522861481, 0.0175931453704834, 0.12186291813850403, 0.050128500908613205, 0.0007250979542732239, -0.11926645040512085, 0.09062512964010239, 0.029430991038680077, -0.05209023877978325, 0.04674858972430229, 0.07807627320289612, 0.20056791603565216, 0.08357866108417511, 0.06494186073541641, 0.05303197726607323, -0.0949070006608963, -0.13216345012187958, 0.02095859684050083, 0.035429954528808594, 0.013220532797276974, 0.026818275451660156, 0.1279418170452118, -0.023170359432697296, 0.055000171065330505, -0.008026834577322006, 0.012535826303064823, -0.19789190590381622, -0.1645161509513855, -0.04297390952706337, -0.06835578382015228, 0.014683719724416733, 0.01553224865347147, 0.024983864277601242, 0.05192544683814049, 0.05425817146897316, -0.055427588522434235, 0.05165494233369827, -0.07665993273258209, -0.015841737389564514, 0.016537008807063103, -0.002547574695199728, 0.10228928923606873, -0.020317355170845985, -0.010246377438306808, -0.12696097791194916, 0.01659955270588398, -0.03363679349422455, 0.04786749929189682, -0.018585262820124626, 0.029209978878498077, -0.10234033316373825, -0.06467923521995544, -0.05014568939805031, 0.04604099690914154, 0.02401241473853588, 0.16941873729228973, 0.0038465713150799274, 0.029655752703547478, 0.05143328756093979, 0.09892440587282181, -0.01531639602035284, -0.10144996643066406, -0.05981764569878578, 0.108296699821949, 0.015415756963193417, 0.08272334188222885, -0.009024891071021557, 0.043243370950222015, -0.0821114182472229, 0.3440125286579132, 0.3070217967033386, -0.03368919715285301, 0.032361872494220734, 0.021568577736616135, 0.02668897807598114, 0.08180173486471176, 0.05143052712082863, 0.070295549929142, 0.2886837422847748, -0.09367009252309799, -0.08492930978536606, -0.04351063817739487, 0.022399460896849632, -0.17902646958827972, 0.03946017101407051, -0.049073781818151474, -0.07071860879659653, -0.008856849744915962, 0.08328451961278915, -0.11162284016609192, 0.050819143652915955, 0.11783179640769958, -0.14693520963191986, -0.05198581516742706, -0.02523757517337799, 0.15677529573440552, 0.08882012218236923, 0.08178681135177612, -0.018203945830464363, -0.0795622244477272, 0.08210377395153046, 0.021174173802137375, -0.18392127752304077, -0.14960233867168427, 0.09141328185796738, 0.04542047530412674, 0.22592921555042267, -0.0105033740401268, -0.018152818083763123, 0.11618495732545853, 0.0629216879606247, -0.049699362367391586, 0.03206802159547806, 0.03116493672132492, 0.004314346704632044, -0.0031668427400290966, -0.1378125548362732, -0.010586832650005817, -0.11882185190916061, 0.024147721007466316, -0.040453385561704636, 0.0748591274023056, -0.01751263625919819, -0.03965779393911362, -0.05325644835829735, 0.14552851021289825, -0.09854105114936829, 0.015989942476153374, 0.04718894511461258, 0.021664822474122047, -0.04859914258122444, -0.06030065193772316, -0.019017772749066353, 0.031623970717191696, -0.13771699368953705, -0.1035080999135971, -0.005295957904309034, -0.011772554367780685, 0.033829666674137115, 0.019438326358795166, -0.13017770648002625, -0.022688880562782288, -0.09973709285259247, -0.00447204802185297, -0.13056567311286926, -0.016612287610769272, 0.06023424491286278, 0.043171338737010956, 0.018533457070589066, -0.04603781923651695, 0.04072543606162071, -0.016202710568904877, -0.06814399361610413, -0.06755740940570831 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/roberta-large-bne-capitel-ner # Spanish RoBERTa-large trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset. RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019. Original pre-trained model can be found here: https://huggingface.co/BSC-TeMU/roberta-large-bne ## Dataset The dataset used is the one from the [CAPITEL competition at IberLEF 2020](https://sites.google.com/view/capitel2020) (sub-task 1). ## Evaluation and results F1 Score: 0.8998 For evaluation details visit our [GitHub repository](https://github.com/PlanTL-SANIDAD/lm-spanish). ## Citing Check out our paper for all the details: https://arxiv.org/abs/2107.07253 ``` @misc{gutierrezfandino2021spanish, title={Spanish Language Models}, author={Asier Gutiérrez-Fandiño and Jordi Armengol-Estapé and Marc Pàmies and Joan Llop-Palao and Joaquín Silveira-Ocampo and Casimiro Pio Carrino and Aitor Gonzalez-Agirre and Carme Armentano-Oller and Carlos Rodriguez-Penagos and Marta Villegas}, year={2021}, eprint={2107.07253}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
{"language": ["es"], "license": "apache-2.0", "tags": ["national library of spain", "spanish", "bne", "capitel", "ner"], "datasets": ["bne", "capitel"], "metrics": ["f1"]}
token-classification
BSC-LT/roberta-large-bne-capitel-ner
[ "transformers", "pytorch", "roberta", "token-classification", "national library of spain", "spanish", "bne", "capitel", "ner", "es", "dataset:bne", "dataset:capitel", "arxiv:1907.11692", "arxiv:2107.07253", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1907.11692", "2107.07253" ]
[ "es" ]
TAGS #transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #ner #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL # Spanish RoBERTa-large trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset. RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019. Original pre-trained model can be found here: URL ## Dataset The dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 1). ## Evaluation and results F1 Score: 0.8998 For evaluation details visit our GitHub repository. ## Citing Check out our paper for all the details: URL
[ "# Spanish RoBERTa-large trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset.\nRoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 1).", "## Evaluation and results\nF1 Score: 0.8998\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ "TAGS\n#transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #ner #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Spanish RoBERTa-large trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset.\nRoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 1).", "## Evaluation and results\nF1 Score: 0.8998\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ 95, 144, 29, 24, 13 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #ner #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Spanish RoBERTa-large trained on BNE finetuned for CAPITEL Named Entity Recognition (NER) dataset.\nRoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 1).## Evaluation and results\nF1 Score: 0.8998\n\nFor evaluation details visit our GitHub repository.## Citing \nCheck out our paper for all the details: URL" ]
[ -0.11645864695310593, 0.2342580109834671, -0.006054691504687071, 0.05035162717103958, 0.03008999489247799, -0.02738584578037262, 0.007300427183508873, 0.10042601823806763, -0.05471622571349144, 0.09319423884153366, 0.005281373392790556, 0.1150154247879982, 0.08746612817049026, 0.05966448783874512, 0.005758766550570726, -0.1811681091785431, 0.04464897885918617, -0.01831105910241604, 0.030614705756306648, 0.06145654618740082, 0.0895112082362175, -0.04604262486100197, 0.01386773120611906, 0.012138963676989079, -0.03302818536758423, 0.09058237075805664, -0.022918518632650375, -0.17010754346847534, 0.08731220662593842, 0.052737634629011154, 0.11342697590589523, 0.02512206695973873, -0.012031573802232742, -0.04126717522740364, -0.0027189801912754774, 0.0262399110943079, -0.025290312245488167, 0.055867183953523636, 0.12411810457706451, -0.12659700214862823, 0.21834690868854523, -0.03854271024465561, 0.007154710590839386, 0.03989648073911667, -0.18062451481819153, -0.25942307710647583, -0.14054302871227264, 0.01193090807646513, 0.034763678908348083, 0.07212579250335693, -0.0267848689109087, 0.10263046622276306, -0.06462179124355316, -0.00985208060592413, 0.062333498150110245, -0.1517016589641571, -0.05999834090471268, 0.04319815710186958, 0.02428400330245495, 0.10271871834993362, -0.04859551414847374, 0.0075887334533035755, 0.08264025300741196, 0.011822052299976349, -0.03537904843688011, -0.08600538223981857, -0.13728156685829163, -0.007552643772214651, -0.11069077998399734, -0.09236820787191391, 0.07462529838085175, -0.0660010352730751, -0.07531120628118515, -0.022082971408963203, -0.04729482904076576, 0.043636906892061234, 0.015515058301389217, -0.0077229333110153675, 0.06620027124881744, -0.03208249807357788, 0.050892140716314316, -0.03156450018286705, -0.06559469550848007, -0.015929970890283585, -0.1313723474740982, 0.1576663851737976, 0.00489816814661026, 0.028080323711037636, 0.012878176756203175, 0.056385498493909836, -0.018268397077918053, -0.09441636502742767, -0.06593523919582367, 0.0019981495570391417, -0.022656451910734177, -0.027911385521292686, -0.023046035319566727, -0.11358337104320526, 0.03557805344462395, 0.08817639946937561, -0.18569736182689667, -0.023666610941290855, -0.03966311365365982, 0.013458075933158398, 0.11122957617044449, 0.1716700792312622, -0.03895541653037071, -0.08829885721206665, 0.009782231412827969, -0.0660044476389885, 0.03775686025619507, 0.03296641632914543, -0.0620340034365654, -0.08529787510633469, -0.0015858480473980308, 0.0919518694281578, 0.0782075747847557, -0.0780666172504425, -0.05224916338920593, -0.043276213109493256, 0.155878484249115, -0.09031593054533005, -0.0033253603614866734, -0.017145322635769844, -0.10893577337265015, 0.0192731861025095, -0.0354284830391407, -0.07152199000120163, -0.08199223130941391, -0.04949367418885231, -0.08040457218885422, -0.10642649233341217, -0.07651180773973465, -0.09719441086053848, 0.09225720167160034, -0.13083454966545105, 0.004135277587920427, -0.11767345666885376, -0.07055869698524475, -0.11572705954313278, -0.003965780138969421, -0.10520275682210922, 0.07141043245792389, -0.06321761757135391, 0.065864697098732, 0.005224139429628849, -0.025743532925844193, 0.026509137824177742, -0.038073234260082245, 0.0720677524805069, 0.060597069561481476, 0.06606488674879074, -0.10120463371276855, 0.003916481975466013, -0.1633049100637436, 0.006511094514280558, -0.2452043890953064, 0.11094055324792862, -0.09521377086639404, -0.0027785953134298325, -0.1768486201763153, -0.060354575514793396, -0.03354961425065994, 0.05996120348572731, 0.061547763645648956, 0.1049472838640213, -0.05222021043300629, -0.09333942830562592, 0.19775348901748657, -0.0968940481543541, -0.0382588692009449, 0.08652281016111374, 0.000355978321749717, 0.08297352492809296, 0.11088398098945618, 0.12362291663885117, 0.10169666260480881, -0.01794082298874855, -0.10918685793876648, 0.00969370361417532, 0.029103755950927734, -0.09419938176870346, 0.10731524974107742, -0.10763830691576004, 0.040076691657304764, 0.01346649881452322, -0.12899149954319, 0.006649376358836889, 0.0018072002567350864, -0.01253435481339693, 0.062245894223451614, 0.02113249897956848, -0.06520719081163406, 0.04640989005565643, 0.032969165593385696, -0.02388782799243927, -0.02530391700565815, 0.046594440937042236, 0.04405321925878525, 0.037280961871147156, 0.0057050082832574844, -0.0523911751806736, 0.15227660536766052, -0.13558842241764069, 0.0053216698579490185, -0.17124293744564056, 0.0535348504781723, 0.010274752974510193, -0.04693393036723137, 0.046212658286094666, -0.013848557136952877, 0.0077064624056220055, -0.027278538793325424, -0.016123782843351364, -0.004383178893476725, -0.03562599793076515, -0.07081009447574615, 0.007896999828517437, -0.12176413089036942, 0.006536159664392471, -0.05138677731156349, 0.027933696284890175, -0.09446249157190323, -0.05443378537893295, 0.06119080260396004, 0.05048888921737671, -0.017370639368891716, -0.033903054893016815, 0.019380832090973854, 0.08387032896280289, -0.04255015775561333, -0.05067825689911842, 0.004487316124141216, 0.006107100751250982, -0.03208064287900925, 0.08702217042446136, -0.01690906099975109, -0.06299786269664764, 0.08834119886159897, 0.006784598808735609, -0.014139225706458092, 0.032417699694633484, -0.02635744959115982, 0.055855073034763336, -0.0003541341284289956, -0.005260280333459377, 0.19310471415519714, 0.009830488823354244, 0.09174130856990814, -0.1872134655714035, -0.04318584129214287, 0.015389819629490376, -0.03406718745827675, -0.08415891230106354, 0.18059797585010529, -0.004536382853984833, -0.16801440715789795, 0.0970161110162735, 0.00020201313600409776, -0.025126313790678978, 0.22707073390483856, -0.003903456963598728, -0.09386533498764038, 0.008756062015891075, 0.01936342939734459, 0.03875558823347092, 0.1077384427189827, -0.08895585685968399, -0.011631941422820091, 0.026231933385133743, 0.012087102048099041, 0.09150278568267822, -0.08140672743320465, 0.029695602133870125, -0.0081370510160923, -0.059802912175655365, -0.03509455546736717, 0.019434137269854546, -0.012342771515250206, 0.09267240762710571, 0.0983501598238945, -0.02221018448472023, -0.055793266743421555, -0.03066295199096203, -0.07239460945129395, 0.19471873342990875, -0.12512825429439545, -0.22975632548332214, -0.1999961882829666, 0.029714297503232956, -0.026822976768016815, 0.09786798059940338, -0.0077685024589300156, -0.08803610503673553, -0.022527651861310005, 0.02247954159975052, 0.16389232873916626, -0.061768677085638046, -0.09422747045755386, -0.07939837127923965, 0.03896823897957802, -0.0859406441450119, -0.16747444868087769, 0.02899647131562233, -0.012801551260054111, -0.10172545164823532, -0.019879668951034546, -0.09881167113780975, 0.12562377750873566, 0.06987852603197098, 0.02325621247291565, -0.02889649011194706, -0.03247122839093208, 0.16152895987033844, -0.16646957397460938, -0.016332263126969337, 0.1480662226676941, 0.09855302423238754, -0.029890023171901703, 0.04054723307490349, -0.005984923802316189, -0.046085651963949203, -0.04323233664035797, 0.02977174147963524, -0.060319364070892334, -0.2798369526863098, -0.11012262850999832, -0.03535931929945946, -0.05230281129479408, 0.060168273746967316, 0.029839705675840378, -0.008731960318982601, 0.08731304109096527, -0.03519883751869202, -0.02686546929180622, -0.021647859364748, 0.047173209488391876, 0.1095513179898262, 0.020584823563694954, 0.02458934858441353, -0.0764203816652298, -0.0439276285469532, 0.16963842511177063, 0.06930481642484665, 0.08973126858472824, -0.040938448160886765, 0.15971925854682922, 0.06691181659698486, 0.06075118109583855, -0.017561305314302444, 0.07050839811563492, -0.007185301743447781, 0.026585277169942856, -0.04647766798734665, -0.0576714426279068, -0.09919837862253189, -0.03493715450167656, -0.00821713637560606, 0.007220122497528791, -0.07709084451198578, -0.1734105795621872, 0.04578709229826927, 0.12876786291599274, 0.03690610080957413, -0.21818672120571136, -0.0726814940571785, 0.04561396688222885, -0.03271326422691345, -0.10802821815013885, -0.004900727886706591, 0.01663806475698948, -0.1167738139629364, 0.08320064097642899, 0.028260206803679466, 0.0871783122420311, -0.12638001143932343, -0.018378492444753647, -0.12224285304546356, 0.028026333078742027, -0.01612296886742115, 0.0927608385682106, -0.10124414414167404, 0.22010654211044312, 0.034412749111652374, 0.06697733700275421, -0.03194230794906616, 0.023898186162114143, -0.017703309655189514, -0.012134344317018986, 0.13516445457935333, 0.027373066172003746, -0.0046559954062104225, -0.05226393789052963, -0.035028617829084396, 0.03980838507413864, -0.020853539928793907, -0.10798322409391403, 0.1031159982085228, 0.03360135108232498, 0.001615611370652914, -0.08991562575101852, -0.08658439666032791, -0.008537176996469498, -0.11169765889644623, 0.005216692108660936, -0.08382957428693771, 0.013057311065495014, -0.04554826021194458, -0.03796729817986488, -0.012642660178244114, 0.10626719892024994, -0.06572136282920837, -0.07811668515205383, -0.12612679600715637, 0.06469884514808655, 0.15517295897006989, -0.07640867680311203, 0.01451308187097311, -0.0529850572347641, 0.0385611467063427, -0.03326394781470299, -0.054245613515377045, 0.057513199746608734, -0.11365262418985367, -0.061549536883831024, -0.018627503886818886, 0.06730595976114273, 0.09013019502162933, 0.015007821843028069, 0.06384115666151047, 0.0297919362783432, 0.003518178593367338, -0.07431843876838684, -0.01692597195506096, -0.03486752137541771, 0.0634593665599823, 0.10951841622591019, -0.018692707642912865, -0.16221778094768524, -0.0296462494879961, -0.003368911100551486, 0.04679643362760544, 0.07861915975809097, 0.005245414096862078, 0.07386858016252518, 0.25284460186958313, -0.12006749957799911, -0.16846372187137604, 0.0037420939188450575, 0.06853251904249191, 0.03740466386079788, -0.07279107719659805, -0.23101305961608887, 0.022743085399270058, 0.16469532251358032, -0.03422561660408974, 0.017165115103125572, -0.2935198247432709, -0.02442622184753418, 0.07601075619459152, 0.05441049486398697, 0.11836057156324387, -0.07902086526155472, -0.09859815984964371, -0.026620274409651756, -0.021746965125203133, 0.06961756944656372, -0.032355181872844696, 0.036703385412693024, -0.02637558802962303, -0.04467401280999184, 0.04474097862839699, 0.0003696660860441625, 0.15529116988182068, -0.041431255638599396, 0.010552962310612202, 0.006073566619306803, 0.051904790103435516, 0.14845387637615204, 0.01332809403538704, 0.06040942296385765, 0.12654848396778107, 0.022482845932245255, -0.20280209183692932, -0.03922073543071747, -0.08860956132411957, 0.08006496727466583, -0.0580146387219429, -0.00964102242141962, -0.060035496950149536, 0.07438940554857254, 0.01772034540772438, -0.0011628407519310713, 0.10259728878736496, -0.06632097065448761, 0.10973715037107468, 0.10090424120426178, 0.14967080950737, 0.08104093372821808, 0.0021866108290851116, 0.042317185550928116, 0.004531902261078358, 0.07277118414640427, -0.03680160641670227, 0.012070896103978157, 0.11028755456209183, 0.005471035372465849, 0.0028386928606778383, -0.002082953928038478, -0.1836473047733307, 0.03749779239296913, 0.1670423299074173, 0.00394871411845088, 0.005329969804733992, -0.03948517516255379, -0.08727159351110458, -0.0010237764799967408, -0.006630073767155409, 0.16650457680225372, 0.04747340455651283, -0.08875639736652374, -0.008419082500040531, -0.0012692561140283942, -0.01733838953077793, 0.14338240027427673, 0.0007807998335920274, 0.013156101107597351, -0.09730060398578644, 0.09520035237073898, 0.11763624101877213, -0.11202776432037354, -0.0059720855206251144, 0.08884118497371674, -0.025390787050127983, -0.03926237300038338, 0.013663070276379585, 0.08691708743572235, -0.28225070238113403, -0.09510289877653122, -0.10476218909025192, -0.08567479997873306, 0.007469154428690672, 0.20261643826961517, 0.028317375108599663, 0.03523941710591316, 0.026872148737311363, 0.023792000487446785, -0.006785149686038494, -0.011837237514555454, 0.013932883739471436, -0.004857956897467375, 0.013932068832218647, 0.041833944618701935, 0.000667628541123122, -0.042108938097953796, -0.023298420011997223, -0.003639650996774435, -0.14542561769485474, 0.01654565893113613, -0.12856994569301605, 0.03364107757806778, -0.056861761957407, -0.01218460500240326, -0.0512440986931324, 0.00189111172221601, -0.029291018843650818, -0.048020731657743454, -0.058442872017621994, 0.014590746723115444, -0.02912995219230652, 0.09727403521537781, -0.10407455265522003, 0.03129331022500992, 0.007754435762763023, -0.03890139237046242, 0.043585404753685, 0.013035223819315434, 0.0658215880393982, 0.11189322918653488, -0.12974503636360168, 0.03989173844456673, 0.033800434321165085, 0.0185762420296669, 0.0481504462659359, -0.04693744704127312, 0.05360627919435501, 0.06668764352798462, -0.02564428187906742, 0.04861504212021828, -0.04547746852040291, -0.10788142681121826, -0.0010848314268514514, 0.06028660759329796, -0.05800578370690346, -0.013421567156910896, 0.14535878598690033, 0.14412564039230347, 0.030204759910702705, 0.042459696531295776, -0.07009334117174149, -0.023940129205584526, -0.10784640908241272, -0.0148946363478899, -0.043221015483140945, 0.00632781907916069, -0.05567547678947449, 0.02073359675705433, 0.04006125405430794, 0.03559036925435066, 0.15064875781536102, 0.09203622490167618, 0.13759247958660126, 0.015660125762224197, -0.0034106879029423, -0.006467696279287338, 0.0009699578513391316, 0.1161675676703453, 0.10273759812116623, 0.005839914083480835, 0.005854117684066296, 0.06730258464813232, -0.006007079966366291, -0.029131731018424034, 0.030046720057725906, 0.16875137388706207, 0.22147758305072784, 0.10504240542650223, 0.040789373219013214, -0.014962957240641117, 0.037374429404735565, 0.017070837318897247, 0.016715386882424355, -0.005130787845700979, 0.04209189489483833, -0.006514452863484621, 0.14118625223636627, -0.1528048813343048, 0.09800981730222702, 0.07060413062572479, 0.00765571603551507, -0.08090061694383621, -0.15246722102165222, -0.01979079842567444, -0.053947944194078445, 0.014039287343621254, -0.13863717019557953, 0.043590743094682693, -0.0395154170691967, 0.03231441229581833, -0.0785866379737854, 0.024738555774092674, -0.03424050658941269, -0.17985647916793823, 0.1090705469250679, 0.040661364793777466, 0.17216713726520538, 0.003677502740174532, 0.05351944640278816, 0.0017649286892265081, 0.15124164521694183, 0.05687427520751953, 0.07409881055355072, 0.02440802939236164, 0.05038873478770256, -0.03383419290184975, -0.03538164868950844, 0.009659146890044212, 0.041812676936388016, 0.05174444243311882, 0.20857886970043182, 0.035039085894823074, -0.03602242469787598, 0.009537585079669952, 0.2467571645975113, -0.014533504843711853, 0.05237562954425812, -0.08548186719417572, 0.22086495161056519, 0.02680191583931446, 0.05043205991387367, 0.016465038061141968, -0.12782122194766998, -0.03650592267513275, 0.15856316685676575, 0.13518738746643066, 0.060200415551662445, -0.044979438185691833, 0.01453491672873497, 0.005422369576990604, 0.07988497614860535, 0.06487945467233658, -0.01548831257969141, 0.3720771074295044, -0.05458873510360718, -0.0008572782971896231, -0.041094180196523666, 0.10156988352537155, -0.005285481456667185, 0.11766872555017471, -0.03984333574771881, -0.06067044287919998, -0.02131788618862629, 0.1518314629793167, -0.11233887076377869, -0.3145532011985779, 0.08371715247631073, -0.11169387400150299, -0.15295636653900146, -0.0642535537481308, -0.009463248774409294, 0.07480136305093765, 0.08440430462360382, 0.06314466893672943, -0.05206849053502083, 0.0029527025762945414, 0.04182891920208931, -0.09386445581912994, -0.20934852957725525, 0.08776015788316727, 0.016129644587635994, 0.2619565427303314, -0.022050682455301285, -0.012721613049507141, 0.09093936532735825, 0.005482438486069441, -0.09645107388496399, 0.04282733425498009, 0.02501767687499523, -0.029166465625166893, 0.08373826742172241, 0.022927863523364067, 0.026860982179641724, 0.04796728491783142, 0.07510445266962051, -0.06892941892147064, 0.03630143404006958, 0.060669805854558945, -0.010646120645105839, -0.12734858691692352, 0.10031556338071823, -0.13498848676681519, 0.0689854621887207, 0.1845422387123108, -0.0015458930283784866, 0.044899217784404755, -0.0711001381278038, 0.09665278345346451, 0.018325330689549446, 0.12508374452590942, -0.018445266410708427, -0.14572350680828094, -0.016188260167837143, -0.03210977837443352, -0.023917417973279953, -0.22454875707626343, 0.10039172321557999, 0.02676602266728878, -0.023374170064926147, -0.03065345250070095, 0.0511273629963398, -0.05991456285119057, 0.05773676186800003, 0.015923652797937393, -0.018997592851519585, -0.003927985206246376, 0.03520234301686287, -0.07493699342012405, -0.024720817804336548 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/roberta-large-bne-capitel-pos # Spanish RoBERTa-large trained on BNE finetuned for CAPITEL Part of Speech (POS) dataset RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019. Original pre-trained model can be found here: https://huggingface.co/BSC-TeMU/roberta-large-bne ## Dataset The dataset used is the one from the [CAPITEL competition at IberLEF 2020](https://sites.google.com/view/capitel2020) (sub-task 2). ## Evaluation and results F1 Score: 0.9851 (average of 5 runs). For evaluation details visit our [GitHub repository](https://github.com/PlanTL-SANIDAD/lm-spanish). ## Citing Check out our paper for all the details: https://arxiv.org/abs/2107.07253 ``` @misc{gutierrezfandino2021spanish, title={Spanish Language Models}, author={Asier Gutiérrez-Fandiño and Jordi Armengol-Estapé and Marc Pàmies and Joan Llop-Palao and Joaquín Silveira-Ocampo and Casimiro Pio Carrino and Aitor Gonzalez-Agirre and Carme Armentano-Oller and Carlos Rodriguez-Penagos and Marta Villegas}, year={2021}, eprint={2107.07253}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
{"language": ["es"], "license": "apache-2.0", "tags": ["national library of spain", "spanish", "bne", "capitel", "pos"], "datasets": ["bne", "capitel"], "metrics": ["f1"], "widget": [{"text": "Festival de San Sebasti\u00e1n: Johnny Depp recibir\u00e1 el premio Donostia en pleno rifirrafe judicial con Amber Heard"}, {"text": "El alcalde de Vigo, Abel Caballero, ha comenzado a colocar las luces de Navidad en agosto."}, {"text": "Gracias a los datos de la BNE, se ha podido lograr este modelo del lenguaje."}, {"text": "El Tribunal Superior de Justicia se pronunci\u00f3 ayer: \"Hay base legal dentro del marco jur\u00eddico actual\"."}]}
token-classification
BSC-LT/roberta-large-bne-capitel-pos
[ "transformers", "pytorch", "roberta", "token-classification", "national library of spain", "spanish", "bne", "capitel", "pos", "es", "dataset:bne", "dataset:capitel", "arxiv:1907.11692", "arxiv:2107.07253", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1907.11692", "2107.07253" ]
[ "es" ]
TAGS #transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #pos #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL # Spanish RoBERTa-large trained on BNE finetuned for CAPITEL Part of Speech (POS) dataset RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019. Original pre-trained model can be found here: URL ## Dataset The dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 2). ## Evaluation and results F1 Score: 0.9851 (average of 5 runs). For evaluation details visit our GitHub repository. ## Citing Check out our paper for all the details: URL
[ "# Spanish RoBERTa-large trained on BNE finetuned for CAPITEL Part of Speech (POS) dataset\nRoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 2).", "## Evaluation and results\nF1 Score: 0.9851 (average of 5 runs).\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ "TAGS\n#transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #pos #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Spanish RoBERTa-large trained on BNE finetuned for CAPITEL Part of Speech (POS) dataset\nRoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 2).", "## Evaluation and results\nF1 Score: 0.9851 (average of 5 runs).\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ 95, 139, 29, 32, 13 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #token-classification #national library of spain #spanish #bne #capitel #pos #es #dataset-bne #dataset-capitel #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Spanish RoBERTa-large trained on BNE finetuned for CAPITEL Part of Speech (POS) dataset\nRoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL## Dataset\nThe dataset used is the one from the CAPITEL competition at IberLEF 2020 (sub-task 2).## Evaluation and results\nF1 Score: 0.9851 (average of 5 runs).\n\nFor evaluation details visit our GitHub repository.## Citing \nCheck out our paper for all the details: URL" ]
[ -0.13967950642108917, 0.23091740906238556, -0.004663147032260895, 0.03936063498258591, 0.051669891923666, -0.033548060804605484, -0.008802459575235844, 0.11279455572366714, -0.02828102372586727, 0.08958760648965836, 0.018612738698720932, 0.0924404188990593, 0.09174397587776184, 0.08035590499639511, 0.024931492283940315, -0.18725280463695526, 0.0503598116338253, -0.04055880010128021, -0.009208965115249157, 0.054904863238334656, 0.10147040337324142, -0.054936643689870834, 0.00923078041523695, 0.009219476953148842, -0.057613518089056015, 0.08436291664838791, -0.03581807389855385, -0.16409669816493988, 0.08691666275262833, 0.020278766751289368, 0.08907056599855423, 0.027384107932448387, 0.01642933115363121, -0.012652905657887459, -0.003834426635876298, 0.0417557917535305, 0.013387581333518028, 0.043636027723550797, 0.12708629667758942, -0.10726303607225418, 0.19762487709522247, -0.027847973629832268, 0.0019979362841695547, 0.060318633913993835, -0.18634100258350372, -0.22303597629070282, -0.13388073444366455, -0.028461026027798653, 0.07006336003541946, 0.09450387209653854, -0.02861638367176056, 0.06933949142694473, -0.0488877072930336, -0.02497829496860504, 0.057664986699819565, -0.16535139083862305, -0.05682375654578209, 0.0662219375371933, 0.0256305243819952, 0.09382078051567078, -0.04691531881690025, 0.016347207129001617, 0.08996749669313431, 0.006125529762357473, -0.0431581512093544, -0.0642048642039299, -0.11436015367507935, -0.017664549872279167, -0.11518389731645584, -0.08939176052808762, 0.1201198548078537, -0.0714033767580986, -0.0835331529378891, -0.03593100234866142, -0.056894488632678986, 0.003204901935532689, 0.00804182980209589, -0.014218613505363464, 0.08101650327444077, -0.002420786302536726, 0.04649897664785385, -0.0476512610912323, -0.07906220853328705, -0.013755534775555134, -0.1025828868150711, 0.1530858278274536, 0.006868423894047737, 0.026210488751530647, 0.008433511480689049, 0.061906494200229645, -0.016248861327767372, -0.08550338447093964, -0.06711946427822113, -0.01053908932954073, -0.03159891068935394, -0.027938149869441986, -0.04229554906487465, -0.13955873250961304, 0.03650955483317375, 0.06929069757461548, -0.17518573999404907, -0.0340188704431057, -0.03246626630425453, -0.004318689927458763, 0.12636230885982513, 0.1611529439687729, -0.04463072493672371, -0.0954524353146553, 0.0034537375904619694, -0.0582919605076313, 0.016317619010806084, 0.0339287593960762, -0.054949816316366196, -0.07082725316286087, -0.027998918667435646, 0.08721896260976791, 0.07558770477771759, -0.06635873764753342, -0.044716283679008484, -0.05658775195479393, 0.15089663863182068, -0.08722557127475739, -0.024897512048482895, -0.029582148417830467, -0.11252402514219284, 0.016221143305301666, -0.04896074905991554, -0.03819822147488594, -0.07170585542917252, -0.06987378001213074, -0.0812043845653534, -0.11627202481031418, -0.07748428732156754, -0.08171242475509644, 0.07244646549224854, -0.15437117218971252, -0.004142455291002989, -0.10333273559808731, -0.09832045435905457, -0.12675201892852783, -0.004108046647161245, -0.10399145632982254, 0.0557137094438076, -0.0560387521982193, 0.057257279753685, -0.0012733573094010353, -0.023572519421577454, 0.029583407565951347, -0.03157045692205429, 0.06256236881017685, 0.06119690462946892, 0.07069995254278183, -0.07992237061262131, -0.0009169927798211575, -0.16509629786014557, -0.00804390199482441, -0.2310985028743744, 0.12430063635110855, -0.08382508903741837, 0.030240708962082863, -0.17817476391792297, -0.04644498601555824, -0.06982925534248352, 0.07365591824054718, 0.06598157435655594, 0.12255896627902985, -0.06658080965280533, -0.08142755180597305, 0.22617466747760773, -0.09227088838815689, -0.04653730243444443, 0.08952836692333221, 0.016688209027051926, 0.09026698023080826, 0.10994873195886612, 0.11325088143348694, 0.05002432316541672, -0.034979961812496185, -0.09487246721982956, -0.028356967493891716, 0.03745950013399124, -0.10675650835037231, 0.10654272884130478, -0.13994358479976654, 0.032129112631082535, 0.01737496443092823, -0.1296067237854004, -0.000033618154702708125, -0.00244291964918375, -0.014572352170944214, 0.046429529786109924, -0.0059618293307721615, -0.08034984767436981, 0.040391188114881516, 0.04236424341797829, -0.031977009028196335, -0.028130952268838882, -0.0015998712042346597, 0.05285913497209549, 0.035516172647476196, -0.004616941791027784, -0.044406525790691376, 0.12999017536640167, -0.13089962303638458, 0.0008767654071561992, -0.16309216618537903, 0.04636114835739136, 0.009279790334403515, -0.017298394814133644, 0.029800962656736374, 0.01920735277235508, 0.008132330141961575, -0.028559399768710136, -0.005429296754300594, 0.0022349394857883453, -0.019145678728818893, -0.07056985050439835, -0.010200903750956059, -0.10344082117080688, 0.010132129304111004, -0.05773213133215904, 0.049230169504880905, -0.07291629165410995, -0.04537639766931534, 0.03032900206744671, 0.06123385950922966, -0.029031410813331604, -0.062243711203336716, 0.03263959288597107, 0.0595344714820385, -0.0284122247248888, -0.07742591947317123, 0.007140026427805424, 0.022410502657294273, 0.02476501651108265, 0.07837267220020294, -0.03230105713009834, -0.07100330293178558, 0.09646563977003098, 0.02064298838376999, -0.011633639223873615, 0.0478535071015358, -0.04069948196411133, 0.06063539534807205, -0.024627922102808952, -0.01153575535863638, 0.19050660729408264, -0.003743893699720502, 0.09529247879981995, -0.1720743179321289, -0.027038458734750748, 0.01716449297964573, -0.021384572610259056, -0.0814053937792778, 0.12840919196605682, 0.012973615899682045, -0.1779087781906128, 0.08383183926343918, -0.029831934720277786, 0.014478489756584167, 0.22114354372024536, 0.008708390407264233, -0.10054143518209457, 0.021536564454436302, 0.011627829633653164, 0.008844711817800999, 0.11932593584060669, -0.14494609832763672, -0.03934090584516525, 0.016322117298841476, 0.01470237784087658, 0.106651671230793, -0.07451117038726807, 0.028600923717021942, -0.009733826853334904, -0.05333275720477104, -0.007011660374701023, 0.025103453546762466, -0.030176904052495956, 0.08630473166704178, 0.10492908209562302, -0.018970929086208344, -0.06481201946735382, -0.022717639803886414, -0.06582271307706833, 0.17982150614261627, -0.11236017942428589, -0.2344064861536026, -0.15319475531578064, 0.033007267862558365, -0.03923358768224716, 0.11975382268428802, 0.008424506522715092, -0.10960923135280609, -0.03358905017375946, 0.02875124290585518, 0.17447155714035034, -0.0757216215133667, -0.06954345107078552, -0.03585497662425041, 0.028418904170393944, -0.07617699354887009, -0.16591884195804596, 0.02220122702419758, -0.012116795405745506, -0.12094644457101822, -0.039346128702163696, -0.08060616254806519, 0.1036522164940834, 0.07290622591972351, 0.035786356776952744, -0.01051568053662777, -0.03073199652135372, 0.16834822297096252, -0.15674740076065063, -0.02533293515443802, 0.1666208952665329, 0.13064657151699066, -0.04947201535105705, 0.056252457201480865, -0.004774121101945639, -0.07334857434034348, -0.029832888394594193, 0.04648786038160324, -0.05391710624098778, -0.2645367383956909, -0.11897458136081696, -0.032335273921489716, -0.06929171085357666, 0.08171789348125458, 0.043793439865112305, 0.014769737608730793, 0.09632837027311325, -0.042592838406562805, -0.03712188079953194, -0.03723815456032753, 0.05074269324541092, 0.0542629174888134, 0.0009674246539361775, 0.03591236472129822, -0.05251041054725647, -0.03199907764792442, 0.16075725853443146, 0.05681493133306503, 0.06119859591126442, -0.041104938834905624, 0.13608744740486145, 0.07778064161539078, 0.06758539378643036, -0.03786362707614899, 0.051468804478645325, 0.010664796456694603, 0.02145322784781456, -0.04018900543451309, -0.04468783363699913, -0.10900498926639557, -0.02761116251349449, 0.01361064799129963, -0.016425393521785736, -0.09095745533704758, -0.15003453195095062, 0.0350998155772686, 0.18382404744625092, 0.0630686804652214, -0.2224023938179016, -0.07407762110233307, 0.04180978611111641, -0.038152486085891724, -0.11258244514465332, 0.0168136153370142, 0.004914135206490755, -0.11049310863018036, 0.07334911823272705, 0.024905243888497353, 0.08024393767118454, -0.10765457153320312, -0.006248038727790117, -0.0695401281118393, 0.04018636420369148, -0.00998670980334282, 0.06973419338464737, -0.08911532908678055, 0.22947998344898224, 0.03242677450180054, 0.08417829126119614, -0.04513625055551529, 0.038975175470113754, -0.014561701565980911, -0.03888055309653282, 0.1466289460659027, 0.0273636132478714, -0.0240809116512537, -0.04454374313354492, -0.05300481244921684, 0.03824929893016815, -0.022137699648737907, -0.13481344282627106, 0.10626363754272461, 0.029773328453302383, -0.009928563609719276, -0.08311440050601959, -0.06313572078943253, -0.05447135865688324, -0.09494830667972565, 0.026379479095339775, -0.06263039261102676, 0.0418848991394043, -0.049808237701654434, -0.029480328783392906, -0.037927087396383286, 0.14869557321071625, -0.008035734295845032, -0.07705382257699966, -0.13313615322113037, 0.08706187456846237, 0.15349385142326355, -0.04368755221366882, 0.026023531332612038, -0.04193216189742088, 0.05388852208852768, -0.022021416574716568, -0.0382736474275589, 0.08041111379861832, -0.13131368160247803, -0.05940353125333786, -0.015051928348839283, 0.08872593939304352, 0.08162131160497665, -0.00397925078868866, 0.06533876806497574, 0.035828933119773865, 0.011777087114751339, -0.0651179701089859, 0.01894211210310459, -0.014621538110077381, 0.08122235536575317, 0.09039804339408875, -0.01434169802814722, -0.14239737391471863, -0.03022468276321888, -0.009508741088211536, 0.04272635653614998, 0.08841045200824738, 0.003907002042979002, 0.08013296127319336, 0.2239908128976822, -0.11289016157388687, -0.1525200456380844, 0.016333891078829765, 0.07396198809146881, 0.044561535120010376, -0.0959664061665535, -0.24602562189102173, 0.017205694690346718, 0.1456330418586731, -0.047223832458257675, -0.009067375212907791, -0.33241721987724304, -0.034395843744277954, 0.0866178572177887, 0.03603231534361839, 0.1438770890235901, -0.08140897005796432, -0.08870038390159607, -0.01923072710633278, 0.0046902988106012344, 0.0524543896317482, -0.03918059915304184, 0.04136143624782562, -0.00827735848724842, -0.0629744753241539, 0.05163319408893585, -0.009126079268753529, 0.1588979959487915, -0.039366912096738815, -0.011274187825620174, 0.006467040162533522, 0.08785972744226456, 0.13157036900520325, 0.012934177182614803, 0.08326204866170883, 0.17188136279582977, 0.03625829517841339, -0.20575179159641266, -0.04312462732195854, -0.08441575616598129, 0.0704847201704979, -0.04184100404381752, -0.01835329644382, -0.040659282356500626, 0.0704619362950325, 0.027339396998286247, 0.00827736221253872, 0.07068971544504166, -0.07464320957660675, 0.11294210702180862, 0.09084895253181458, 0.14709527790546417, 0.059145160019397736, -0.0002454530622344464, 0.03766220808029175, 0.007920339703559875, 0.05947120860219002, -0.038947559893131256, 0.007442957255989313, 0.0954384133219719, 0.005914479959756136, -0.013107609935104847, -0.0018589841201901436, -0.16458146274089813, 0.04044928774237633, 0.1573513299226761, -0.004208920523524284, 0.003566404804587364, -0.05158048868179321, -0.07069050520658493, -0.005305461585521698, 0.005654401611536741, 0.176795095205307, 0.039706792682409286, -0.08964559435844421, -0.019304122775793076, 0.011289121583104134, -0.013581515289843082, 0.13467447459697723, -0.03132796660065651, 0.009737229906022549, -0.10828857123851776, 0.12659600377082825, 0.1119358018040657, -0.10374605655670166, -0.0013549088034778833, 0.09518001228570938, -0.02329816110432148, -0.04507189616560936, 0.008804666809737682, 0.04109763726592064, -0.26346197724342346, -0.07716070860624313, -0.09338966012001038, -0.0717347264289856, -0.0007613216293975711, 0.16191516816616058, 0.010196040384471416, 0.029696449637413025, 0.019151873886585236, 0.04116295278072357, -0.026240170001983643, -0.029795292764902115, 0.0358365923166275, -0.003697996260598302, -0.01086321659386158, 0.05483204126358032, -0.0007176391663961112, -0.05053693801164627, -0.025085922330617905, -0.014536475762724876, -0.11060984432697296, 0.012580332346260548, -0.12386251986026764, -0.009540854953229427, -0.06332333385944366, -0.02837604284286499, -0.037452131509780884, -0.013651886954903603, -0.0326608270406723, -0.046243418008089066, -0.06540566682815552, 0.004493765067309141, -0.031243739649653435, 0.0794651061296463, -0.0883576050400734, 0.04080025106668472, -0.010671017691493034, -0.04667842015624046, 0.04724007844924927, 0.03136100992560387, 0.05833769962191582, 0.08309263736009598, -0.10316062718629837, 0.0320766344666481, 0.049420375376939774, 0.003760579274967313, 0.04187385365366936, -0.046488385647535324, 0.04678406938910484, 0.07844826579093933, -0.02147158607840538, 0.03343696519732475, -0.045194171369075775, -0.10517846792936325, -0.021932264789938927, 0.0751725360751152, -0.019279826432466507, -0.01563074253499508, 0.11145687103271484, 0.1761162132024765, 0.035779088735580444, 0.02591116540133953, -0.0693209245800972, -0.023411408066749573, -0.11896856129169464, -0.009190300479531288, -0.0569424107670784, 0.029105868190526962, -0.04795890301465988, 0.035978883504867554, 0.05389971658587456, 0.027040772140026093, 0.16330204904079437, 0.089352086186409, 0.1655891090631485, 0.026300420984625816, 0.006068959832191467, 0.02636501006782055, 0.011525815352797508, 0.12630540132522583, 0.1066545620560646, 0.01793738827109337, 0.014648444950580597, 0.08294598758220673, 0.012151828967034817, -0.02482875995337963, 0.04630832374095917, 0.14789508283138275, 0.2307959347963333, 0.11356332153081894, 0.01314434502273798, -0.05828820914030075, 0.004910818766802549, 0.04429103434085846, -0.0062574418261647224, -0.001437284634448588, 0.022718830034136772, -0.04361732304096222, 0.1320391148328781, -0.1497863531112671, 0.07429998368024826, 0.05439906194806099, 0.00892347376793623, -0.08081823587417603, -0.1324189007282257, -0.005617040675133467, -0.0815102756023407, 0.011516767553985119, -0.1375981718301773, 0.04687941446900368, -0.04480430483818054, 0.036250513046979904, -0.07487022131681442, 0.034156568348407745, -0.03250446543097496, -0.18082846701145172, 0.10385974496603012, 0.038096752017736435, 0.19702716171741486, -0.004294478800147772, 0.032611947506666183, 0.0003826423780992627, 0.15560370683670044, 0.057204775512218475, 0.05084767937660217, 0.008689302019774914, 0.053412701934576035, -0.021167179569602013, -0.03660351783037186, 0.0050258091650903225, 0.031017201021313667, 0.06090037524700165, 0.2048773318529129, 0.0443490594625473, -0.03569568321108818, 0.013784964568912983, 0.26611506938934326, -0.030356960371136665, 0.06274619698524475, -0.11363979429006577, 0.1720089465379715, 0.02799888513982296, 0.03605291619896889, 0.013389958068728447, -0.14019648730754852, -0.04750863462686539, 0.17591935396194458, 0.12786123156547546, 0.05854783579707146, -0.05476364865899086, 0.015008325688540936, -0.0052344840951263905, 0.06785911321640015, 0.062391892075538635, -0.029838524758815765, 0.36351677775382996, -0.06053952872753143, -0.0086734713986516, -0.009403522126376629, 0.07801178097724915, -0.022301917895674706, 0.13377715647220612, -0.05359999090433121, -0.050003454089164734, -0.008051484823226929, 0.16142427921295166, -0.09223710745573044, -0.3052820563316345, 0.05537581816315651, -0.10043731331825256, -0.15848737955093384, -0.04934815317392349, -0.013428153470158577, 0.10287951678037643, 0.09948038309812546, 0.06275179982185364, -0.03333970159292221, 0.037675485014915466, 0.036418914794921875, -0.0890658050775528, -0.22425396740436554, 0.08573541790246964, 0.007545880042016506, 0.2800644040107727, -0.028322728350758553, -0.037086427211761475, 0.08108323067426682, 0.0013322901213541627, -0.12593616545200348, -0.002416631206870079, 0.008501629345119, -0.03901045769453049, 0.09282577782869339, 0.04960542544722557, 0.00946546159684658, 0.05414479598402977, 0.07067596167325974, -0.08698990941047668, 0.037046462297439575, 0.0875215083360672, 0.0005501697887666523, -0.12286727875471115, 0.10224014520645142, -0.12501083314418793, 0.08582434803247452, 0.16166618466377258, -0.010351534932851791, 0.043211378157138824, -0.06104219704866409, 0.11199933290481567, 0.007557301316410303, 0.12264745682477951, -0.022945474833250046, -0.1943376362323761, -0.022024068981409073, -0.023256517946720123, -0.036991678178310394, -0.18295970559120178, 0.09915826469659805, 0.04399869218468666, -0.02418266236782074, -0.02701631933450699, 0.04487184062600136, -0.05623452365398407, 0.05360353738069534, 0.01043175719678402, 0.04620680212974548, -0.007038652431219816, 0.047051746398210526, -0.08945204317569733, -0.04070219025015831 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/roberta-large-bne-sqac # Spanish RoBERTa-large trained on BNE finetuned for Spanish Question Answering Corpus (SQAC) dataset. RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019. Original pre-trained model can be found here: https://huggingface.co/BSC-TeMU/roberta-large-bne ## Dataset The dataset used is the [SQAC corpus](https://huggingface.co/datasets/BSC-TeMU/SQAC). ## Evaluation and results F1 Score: 0.7993 (average of 5 runs). For evaluation details visit our [GitHub repository](https://github.com/PlanTL-SANIDAD/lm-spanish). ## Citing Check out our paper for all the details: https://arxiv.org/abs/2107.07253 ``` @misc{gutierrezfandino2021spanish, title={Spanish Language Models}, author={Asier Gutiérrez-Fandiño and Jordi Armengol-Estapé and Marc Pàmies and Joan Llop-Palao and Joaquín Silveira-Ocampo and Casimiro Pio Carrino and Aitor Gonzalez-Agirre and Carme Armentano-Oller and Carlos Rodriguez-Penagos and Marta Villegas}, year={2021}, eprint={2107.07253}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
{"language": ["es"], "license": "apache-2.0", "tags": ["national library of spain", "spanish", "bne", "qa", "question answering"], "datasets": ["BSC-TeMU/SQAC"], "metrics": ["f1"]}
question-answering
BSC-LT/roberta-large-bne-sqac
[ "transformers", "pytorch", "roberta", "question-answering", "national library of spain", "spanish", "bne", "qa", "question answering", "es", "dataset:BSC-TeMU/SQAC", "arxiv:1907.11692", "arxiv:2107.07253", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1907.11692", "2107.07253" ]
[ "es" ]
TAGS #transformers #pytorch #roberta #question-answering #national library of spain #spanish #bne #qa #question answering #es #dataset-BSC-TeMU/SQAC #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL # Spanish RoBERTa-large trained on BNE finetuned for Spanish Question Answering Corpus (SQAC) dataset. RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019. Original pre-trained model can be found here: URL ## Dataset The dataset used is the SQAC corpus. ## Evaluation and results F1 Score: 0.7993 (average of 5 runs). For evaluation details visit our GitHub repository. ## Citing Check out our paper for all the details: URL
[ "# Spanish RoBERTa-large trained on BNE finetuned for Spanish Question Answering Corpus (SQAC) dataset.\nRoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the SQAC corpus.", "## Evaluation and results\nF1 Score: 0.7993 (average of 5 runs).\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ "TAGS\n#transformers #pytorch #roberta #question-answering #national library of spain #spanish #bne #qa #question answering #es #dataset-BSC-TeMU/SQAC #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #endpoints_compatible #region-us \n", "# Spanish RoBERTa-large trained on BNE finetuned for Spanish Question Answering Corpus (SQAC) dataset.\nRoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL", "## Dataset\nThe dataset used is the SQAC corpus.", "## Evaluation and results\nF1 Score: 0.7993 (average of 5 runs).\n\nFor evaluation details visit our GitHub repository.", "## Citing \nCheck out our paper for all the details: URL" ]
[ 90, 141, 14, 32, 13 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #question-answering #national library of spain #spanish #bne #qa #question answering #es #dataset-BSC-TeMU/SQAC #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #endpoints_compatible #region-us \n# Spanish RoBERTa-large trained on BNE finetuned for Spanish Question Answering Corpus (SQAC) dataset.\nRoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019.\n\nOriginal pre-trained model can be found here: URL## Dataset\nThe dataset used is the SQAC corpus.## Evaluation and results\nF1 Score: 0.7993 (average of 5 runs).\n\nFor evaluation details visit our GitHub repository.## Citing \nCheck out our paper for all the details: URL" ]
[ -0.12058721482753754, 0.20130887627601624, -0.0018756567733362317, 0.010768817737698555, 0.0697709321975708, -0.013132714666426182, 0.012341870926320553, 0.10368715226650238, -0.003790061455219984, 0.07433193176984787, 0.02616170048713684, 0.05112060531973839, 0.0704641342163086, 0.032882146537303925, 0.04707212746143341, -0.19255302846431732, 0.04102198779582977, -0.049118366092443466, 0.0140451081097126, 0.04978098347783089, 0.07592812925577164, -0.0647629126906395, 0.017720306292176247, -0.0038127221632748842, -0.04007568210363388, 0.0704004317522049, -0.06412805616855621, -0.14390847086906433, 0.08038734644651413, 0.025100428611040115, 0.07080627977848053, 0.007101884111762047, 0.031030021607875824, -0.01790623739361763, 0.008299866691231728, 0.05011949688196182, 0.02126363292336464, 0.04118187725543976, 0.10656533390283585, -0.08437146991491318, 0.18412980437278748, -0.01922697015106678, 0.008031954057514668, 0.05123128741979599, -0.17071519792079926, -0.16972380876541138, -0.10533369332551956, -0.018980370834469795, 0.0632786750793457, 0.11976655572652817, -0.02729169838130474, 0.08674080669879913, -0.044614505022764206, -0.02301592379808426, 0.0777335986495018, -0.1656988561153412, -0.05796884000301361, 0.0772869661450386, 0.0853920504450798, 0.12451539188623428, -0.04554951190948486, 0.02524951659142971, 0.09239115566015244, 0.010807869024574757, -0.06012403964996338, -0.0712963417172432, -0.14874587953090668, -0.02761443331837654, -0.12054061889648438, -0.07573188096284866, 0.1324976086616516, -0.08374428004026413, -0.06072990968823433, -0.01806194894015789, -0.06767358630895615, 0.07251355051994324, 0.014269101433455944, 0.01602211222052574, 0.08294294774532318, -0.0030075768008828163, 0.03013400174677372, -0.08729247748851776, -0.07608625292778015, -0.024295689538121223, -0.0955180823802948, 0.17675387859344482, 0.016818366944789886, 0.0336511991918087, -0.05322672054171562, 0.05657678842544556, 0.004167976789176464, -0.09825948625802994, -0.07311296463012695, -0.014542887918651104, -0.0052265748381614685, -0.014112441800534725, -0.054180439561605453, -0.12360624969005585, 0.05504436045885086, 0.09513719379901886, -0.13933756947517395, -0.044808682054281235, -0.0445486456155777, 0.013097082264721394, 0.11609908938407898, 0.17710472643375397, -0.08990121632814407, -0.08561991155147552, 0.02482323721051216, -0.05829400569200516, 0.016398927196860313, 0.03344526141881943, -0.07411803305149078, -0.07550869882106781, -0.07476819306612015, 0.09122854471206665, 0.10324080288410187, -0.045258961617946625, 0.01812203973531723, -0.05986920744180679, 0.12390604615211487, -0.076186902821064, -0.054124750196933746, -0.03755032271146774, -0.10338436812162399, -0.0015186193631961942, -0.045563604682683945, -0.03486844152212143, -0.07829467207193375, -0.08003684133291245, -0.09350121021270752, -0.11388934403657913, -0.08608171343803406, -0.08954638242721558, 0.06444437056779861, -0.16892853379249573, 0.01033133827149868, -0.0981273427605629, -0.12595269083976746, -0.13081732392311096, -0.01183091476559639, -0.10019198060035706, 0.0789615735411644, -0.028717035427689552, 0.06697448343038559, -0.011771523393690586, -0.02852538228034973, 0.04201061651110649, -0.03648870810866356, 0.05381879583001137, 0.02926364168524742, 0.07583174854516983, -0.07974299788475037, 0.003356762696057558, -0.1774950623512268, -0.005421852692961693, -0.18433044850826263, 0.1136738583445549, -0.05108921229839325, 0.0695817768573761, -0.16526475548744202, -0.03773609176278114, -0.07818381488323212, 0.08135803043842316, 0.06645854562520981, 0.12444377690553665, -0.05014852434396744, -0.05464077368378639, 0.23419353365898132, -0.07585527002811432, -0.07685122638940811, 0.08198557049036026, 0.019068922847509384, 0.11987917125225067, 0.10583707690238953, 0.0955103188753128, 0.03409302979707718, -0.027133816853165627, -0.06837683916091919, -0.01411313097923994, 0.06360470503568649, -0.11612624675035477, 0.09831143170595169, -0.13391198217868805, 0.045963969081640244, 0.018353190273046494, -0.15682858228683472, 0.00017633978859521449, 0.0030629034154117107, -0.017771854996681213, 0.04081767052412033, -0.02429826743900776, -0.10548412799835205, 0.026988251134753227, 0.03559814393520355, 0.004664561711251736, -0.016760531812906265, -0.005467844661325216, 0.06136708706617355, 0.03614794462919235, -0.007318310439586639, -0.03090912476181984, 0.09467543661594391, -0.13840648531913757, 0.007819626480340958, -0.179517462849617, 0.010409113951027393, 0.00021575918071903288, 0.008111834526062012, 0.04096873849630356, 0.006579785142093897, 0.016981001943349838, -0.030217915773391724, -0.00023515798966400325, 0.0032485672272741795, -0.02658688649535179, -0.07377563416957855, -0.02256067283451557, -0.08728829771280289, 0.0007683747098781168, -0.06972455978393555, 0.012838771566748619, -0.03784735128283501, -0.03577117249369621, -0.0349678099155426, 0.03403559699654579, -0.03538406640291214, -0.03450335934758186, 0.054443225264549255, 0.038574062287807465, -0.027746427804231644, -0.07227025926113129, 0.01209044735878706, 0.02260911837220192, 0.02257765270769596, 0.09059437364339828, -0.029315199702978134, -0.08997896313667297, 0.09420835971832275, 0.042440686374902725, 0.0073275757022202015, 0.0461258664727211, -0.031486839056015015, 0.046479787677526474, -0.033812861889600754, -0.008404458872973919, 0.21359491348266602, -0.0011749772820621729, 0.07848136872053146, -0.16715863347053528, -0.019972743466496468, 0.0003918255679309368, -0.01869063638150692, -0.059862978756427765, 0.11412658542394638, 0.04716358333826065, -0.17301185429096222, 0.0673697292804718, -0.018541354686021805, -0.005783180706202984, 0.22255876660346985, 0.017348675057291985, -0.11498230695724487, 0.04091094434261322, -0.008588016033172607, -0.0064843036234378815, 0.12045634537935257, -0.14357967674732208, -0.05282091721892357, 0.01612028107047081, 0.0033394608180969954, 0.10671617090702057, -0.06688395887613297, 0.02224479615688324, -0.027512604370713234, -0.03118167631328106, -0.03963286057114601, 0.02253006584942341, -0.03899801895022392, 0.0821637511253357, 0.12926186621189117, 0.01994582824409008, -0.07926812767982483, -0.026621615514159203, -0.07143480330705643, 0.1940591186285019, -0.1002422571182251, -0.21616090834140778, -0.1511191874742508, 0.03606567531824112, -0.03371558338403702, 0.10779383778572083, 0.014296719804406166, -0.12039658427238464, -0.03197956085205078, 0.0333922803401947, 0.16265688836574554, -0.0694742202758789, -0.07707091420888901, -0.041100867092609406, 0.033406812697649, -0.09240603446960449, -0.16341403126716614, 0.03976849094033241, -0.02276933565735817, -0.11945968866348267, -0.050490181893110275, -0.07342582195997238, 0.11659782379865646, 0.07647787779569626, 0.04432031512260437, -0.016039865091443062, -0.022154511883854866, 0.1798134446144104, -0.14080293476581573, -0.020051172003149986, 0.19678641855716705, 0.11026215553283691, -0.04775848239660263, 0.0694635733962059, -0.014132299460470676, -0.07535149157047272, -0.02111057937145233, 0.019956007599830627, -0.06788056343793869, -0.26901519298553467, -0.10486374795436859, -0.048499803990125656, -0.08523158729076385, 0.0870855525135994, 0.04673371836543083, -0.010340358130633831, 0.11165982484817505, -0.07413053512573242, -0.035412807017564774, -0.05310620367527008, 0.06410343199968338, 0.008646597154438496, -0.00882832333445549, 0.021828752011060715, -0.046531129628419876, -0.011206877417862415, 0.13402336835861206, 0.06708026677370071, 0.09289240092039108, -0.04574853926897049, 0.12574204802513123, 0.0790354311466217, 0.03530073165893555, -0.01801540143787861, 0.06201232224702835, -0.009329508058726788, 0.0240594781935215, -0.04058315232396126, -0.05622852221131325, -0.11203861236572266, -0.01488927099853754, -0.030448300763964653, -0.02587447129189968, -0.054274819791316986, -0.1286771446466446, 0.016466520726680756, 0.20076431334018707, 0.06565269082784653, -0.22425949573516846, -0.08877524733543396, 0.042146794497966766, -0.00863842573016882, -0.1219639778137207, 0.034451402723789215, 0.022468432784080505, -0.10042127966880798, 0.03769397363066673, 0.039997801184654236, 0.10845402628183365, -0.10483255237340927, -0.005923762451857328, -0.04975823312997818, 0.048658642917871475, -0.012078657746315002, 0.07040505111217499, -0.08797605335712433, 0.23862577974796295, 0.028587229549884796, 0.0691082701086998, -0.03579963743686676, 0.032573193311691284, -0.007213651202619076, -0.03238655626773834, 0.13936379551887512, 0.025195755064487457, 0.0037353825755417347, 0.003373258514329791, -0.05808015167713165, 0.06192469224333763, -0.028248796239495277, -0.12044210731983185, 0.1119551733136177, 0.011408709920942783, -0.011411339044570923, -0.0789172574877739, -0.08427869528532028, -0.05508935824036598, -0.0950530469417572, 0.013428807258605957, -0.06785495579242706, 0.06047535315155983, -0.04321650043129921, -0.019725831225514412, 0.013822401873767376, 0.13803011178970337, -0.018248528242111206, -0.09428218752145767, -0.11825540661811829, 0.06383771449327469, 0.15772342681884766, -0.04713965952396393, 0.028396327048540115, -0.025646448135375977, 0.05215658247470856, -0.004055608995258808, -0.054815445095300674, 0.07116428762674332, -0.15965139865875244, -0.03805290162563324, -0.01427204255014658, 0.06403520703315735, 0.06848683208227158, -0.003918573725968599, 0.06852654367685318, 0.023774152621626854, -0.03175557404756546, -0.06981796771287918, -0.0025977178011089563, 0.02366105280816555, 0.08796665072441101, 0.10447331517934799, -0.06374713778495789, -0.1223217099905014, -0.02583855763077736, 0.001087833079509437, 0.04393438249826431, 0.03665163740515709, 0.008484759368002415, 0.06364019960165024, 0.2855919599533081, -0.11580361425876617, -0.17729811370372772, 0.012049999088048935, 0.04039997607469559, 0.04522360861301422, -0.0948539674282074, -0.2530372440814972, 0.026834653690457344, 0.11893637478351593, -0.03295100852847099, -0.022014692425727844, -0.3490135669708252, -0.034821007400751114, 0.1001882553100586, 0.03088710643351078, 0.16077394783496857, -0.09442327171564102, -0.08152052015066147, -0.03603531792759895, -0.003853138769045472, 0.011288565583527088, -0.09311036765575409, 0.03545187786221504, -0.009998183697462082, -0.026980483904480934, 0.043401315808296204, -0.018390389159321785, 0.1432269662618637, -0.04141698405146599, -0.018002627417445183, -0.009226682595908642, 0.08632868528366089, 0.16872958838939667, 0.0002305642847204581, 0.07775513082742691, 0.16290399432182312, 0.049692489206790924, -0.16777370870113373, -0.04603021964430809, -0.0807441994547844, 0.04238239675760269, -0.032310884445905685, -0.03911666199564934, -0.027248259633779526, 0.07877059280872345, 0.019343797117471695, 0.0013008840614929795, 0.04316660761833191, -0.08545158803462982, 0.11530032753944397, 0.07075591385364532, 0.16887995600700378, 0.0318806916475296, 0.0033255605958402157, 0.042030561715364456, 0.015296122059226036, 0.08463642001152039, -0.03697484731674194, -0.010399619117379189, 0.08752546459436417, 0.026105277240276337, -0.0211962778121233, -0.0011793308658525348, -0.13870662450790405, 0.03676982223987579, 0.1339937299489975, -0.002103647217154503, -0.04276382923126221, -0.04160988703370094, -0.07862390577793121, -0.050756167620420456, -0.0010317800333723426, 0.16578170657157898, 0.030175909399986267, -0.09141717106103897, -0.016061322763562202, 0.020942874252796173, 0.004615833517163992, 0.12824857234954834, -0.025744134560227394, -0.0003200742357876152, -0.1116485744714737, 0.11113140732049942, 0.1143830418586731, -0.06409384310245514, -0.006921177264302969, 0.11680949479341507, -0.03589228168129921, -0.05351882055401802, 0.0036926409229636192, -0.012334338389337063, -0.27071404457092285, -0.0678635984659195, -0.07814746350049973, -0.050558436661958694, -0.01822858862578869, 0.1464209258556366, -0.0013118734350427985, 0.015272621996700764, 0.021298356354236603, 0.0428452342748642, -0.034594908356666565, -0.02007329650223255, -0.003431262681260705, 0.0016598898218944669, -0.006977943237870932, 0.03130681812763214, 0.010011520236730576, -0.028442099690437317, -0.03456324711441994, -0.02102638967335224, -0.12748612463474274, 0.03506489098072052, -0.19939374923706055, 0.001764887128956616, -0.06796834617853165, -0.03225136920809746, -0.059659894555807114, -0.023548083379864693, -0.05464724078774452, -0.04088171198964119, -0.05421258136630058, 0.008776156231760979, -0.036411549896001816, 0.0805015116930008, -0.08897867798805237, 0.051978226751089096, 0.00503322621807456, -0.03790652006864548, 0.05826748535037041, 0.03631983697414398, 0.03907158225774765, 0.06888674199581146, -0.08930426836013794, 0.022687256336212158, 0.03758667781949043, 0.008080475963652134, 0.051888033747673035, -0.015623651444911957, 0.03383005037903786, 0.05439494922757149, -0.03643735125660896, 0.029073920100927353, -0.04325999692082405, -0.08960912376642227, 0.0012734620831906796, 0.08085671812295914, -0.00116190523840487, -0.01876077800989151, 0.0974712148308754, 0.15718263387680054, 0.04119793698191643, 0.019088424742221832, -0.05676138028502464, -0.006122151389718056, -0.10716850310564041, -0.006908546667546034, -0.049407146871089935, 0.033467523753643036, -0.057768478989601135, 0.005490106996148825, 0.05233236029744148, 0.02636103890836239, 0.17830002307891846, 0.08165547251701355, 0.18330445885658264, 0.023312870413064957, 0.06303656846284866, 0.020189831033349037, 0.0024315679911524057, 0.1171744242310524, 0.1039077565073967, 0.011766157113015652, -0.005341291427612305, 0.09395208209753036, 0.004846789408475161, -0.025084177032113075, 0.0887686088681221, 0.14659270644187927, 0.303476482629776, 0.10332858562469482, 0.040497977286577225, -0.05184223875403404, -0.007235747762024403, 0.006719599477946758, 0.034752704203128815, 0.005384265910834074, 0.005785685032606125, -0.06262130290269852, 0.17299191653728485, -0.13192780315876007, 0.06838822364807129, 0.039514295756816864, 0.01870843768119812, -0.0856042429804802, -0.10986125469207764, 0.0032283468171954155, -0.07382834702730179, -0.005686389282345772, -0.13554219901561737, 0.0263994038105011, -0.036753006279468536, 0.028856711462140083, -0.07332292199134827, 0.05460231751203537, -0.009333032183349133, -0.2040480226278305, 0.06650486588478088, 0.04407360404729843, 0.21794863045215607, -0.01553022675216198, 0.03085598722100258, 0.01987357996404171, 0.1300741583108902, 0.05611562356352806, 0.0466887392103672, 0.015151345171034336, 0.05517604574561119, -0.05361729860305786, -0.032723087817430496, -0.010099726729094982, 0.05958028510212898, 0.05856025218963623, 0.19039635360240936, 0.05530732497572899, -0.04358643665909767, 0.024606429040431976, 0.29057005047798157, -0.03784746676683426, 0.032502513378858566, -0.10276926308870316, 0.14475978910923004, 0.019083622843027115, 0.02430683746933937, 0.0022140159271657467, -0.1326833963394165, -0.059462692588567734, 0.20836040377616882, 0.14417600631713867, 0.024157928302884102, -0.048567503690719604, 0.01729230210185051, -0.0031404439359903336, 0.049792349338531494, 0.06995167583227158, 0.010563917458057404, 0.3971864879131317, -0.05084674060344696, -0.022150740027427673, 0.02175157144665718, 0.05259557068347931, -0.04258936271071434, 0.08701624721288681, -0.054761797189712524, -0.05714428797364235, -0.004426260013133287, 0.16017314791679382, -0.136679545044899, -0.31057578325271606, 0.02616899274289608, -0.08726857602596283, -0.15367597341537476, -0.050685688853263855, -0.02250233106315136, 0.10914239287376404, 0.09981943666934967, 0.05000581964850426, -0.0427299328148365, 0.03845794498920441, 0.031293537467718124, -0.07402987778186798, -0.23324421048164368, 0.11612169444561005, 0.031848300248384476, 0.25708574056625366, -0.038441356271505356, -0.06530850380659103, 0.07076527923345566, 0.00045108143240213394, -0.1663704216480255, -0.02577381394803524, 0.018142573535442352, -0.061246659606695175, 0.09715095907449722, 0.036729663610458374, -0.01604512520134449, 0.029905162751674652, 0.08692066371440887, -0.07327727973461151, 0.028930198401212692, 0.08961790055036545, 0.040917664766311646, -0.13927961885929108, 0.10620459169149399, -0.12886784970760345, 0.11281323432922363, 0.16396191716194153, -0.020171845331788063, 0.028400469571352005, -0.046554867178201675, 0.10218194127082825, 0.008372952230274677, 0.1134229227900505, -0.02156299166381359, -0.1632954329252243, -0.027015725150704384, -0.04572921618819237, -0.02491358481347561, -0.22273072600364685, 0.09939798712730408, 0.05666203796863556, -0.012027801014482975, -0.012881961651146412, 0.03626774996519089, 0.0036746629048138857, 0.05029332637786865, 0.001742666820064187, 0.021769171580672264, -0.0027420474216341972, 0.03437309339642525, -0.1105799451470375, -0.07026073336601257 ]
null
null
transformers
**⚠️NOTICE⚠️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED:** https://huggingface.co/PlanTL-GOB-ES/roberta-large-bne # RoBERTa large trained with data from National Library of Spain (BNE) ## Model Description RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019. ## Training corpora and preprocessing The [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) crawls all .es domains once a year. The training corpus consists of 59TB of WARC files from these crawls, carried out from 2009 to 2019. To obtain a high-quality training corpus, the corpus has been preprocessed with a pipeline of operations, including among the others, sentence splitting, language detection, filtering of bad-formed sentences and deduplication of repetitive contents. During the process document boundaries are kept. This resulted into 2TB of Spanish clean corpus. Further global deduplication among the corpus is applied, resulting into 570GB of text. Some of the statistics of the corpus: | Corpora | Number of documents | Number of tokens | Size (GB) | |---------|---------------------|------------------|-----------| | BNE | 201,080,084 | 135,733,450,668 | 570GB | ## Tokenization and pre-training The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original [RoBERTA](https://arxiv.org/abs/1907.11692) model with a vocabulary size of 50,262 tokens. The RoBERTa-large-bne pre-training consists of a masked language model training that follows the approach employed for the RoBERTa large. The training lasted a total of 96 hours with 32 computing nodes each one with 4 NVIDIA V100 GPUs of 16GB VRAM. ## Evaluation and results For evaluation details visit our [GitHub repository](https://github.com/PlanTL-SANIDAD/lm-spanish). ## Citing Check out our paper for all the details: https://arxiv.org/abs/2107.07253 ``` @misc{gutierrezfandino2021spanish, title={Spanish Language Models}, author={Asier Gutiérrez-Fandiño and Jordi Armengol-Estapé and Marc Pàmies and Joan Llop-Palao and Joaquín Silveira-Ocampo and Casimiro Pio Carrino and Aitor Gonzalez-Agirre and Carme Armentano-Oller and Carlos Rodriguez-Penagos and Marta Villegas}, year={2021}, eprint={2107.07253}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
{"language": ["es"], "license": "apache-2.0", "tags": ["national library of spain", "spanish", "bne"], "datasets": ["bne"], "metrics": ["ppl"], "widget": [{"text": "Este a\u00f1o las campanadas de La Sexta las <mask> Pedroche y Chicote."}, {"text": "El artista Antonio Orozco es un colaborador de La <mask>."}, {"text": "Gracias a los datos de la BNE se ha podido <mask> este modelo del lenguaje."}, {"text": "Hay base legal dentro del marco <mask> actual."}]}
fill-mask
BSC-LT/roberta-large-bne
[ "transformers", "pytorch", "roberta", "fill-mask", "national library of spain", "spanish", "bne", "es", "dataset:bne", "arxiv:1907.11692", "arxiv:2107.07253", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1907.11692", "2107.07253" ]
[ "es" ]
TAGS #transformers #pytorch #roberta #fill-mask #national library of spain #spanish #bne #es #dataset-bne #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
️NOTICE️: THIS MODEL HAS BEEN MOVED TO THE FOLLOWING URL AND WILL SOON BE REMOVED: URL RoBERTa large trained with data from National Library of Spain (BNE) ==================================================================== Model Description ----------------- RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the RoBERTa large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the National Library of Spain (Biblioteca Nacional de España) from 2009 to 2019. Training corpora and preprocessing ---------------------------------- The National Library of Spain (Biblioteca Nacional de España) crawls all .es domains once a year. The training corpus consists of 59TB of WARC files from these crawls, carried out from 2009 to 2019. To obtain a high-quality training corpus, the corpus has been preprocessed with a pipeline of operations, including among the others, sentence splitting, language detection, filtering of bad-formed sentences and deduplication of repetitive contents. During the process document boundaries are kept. This resulted into 2TB of Spanish clean corpus. Further global deduplication among the corpus is applied, resulting into 570GB of text. Some of the statistics of the corpus: Tokenization and pre-training ----------------------------- The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original RoBERTA model with a vocabulary size of 50,262 tokens. The RoBERTa-large-bne pre-training consists of a masked language model training that follows the approach employed for the RoBERTa large. The training lasted a total of 96 hours with 32 computing nodes each one with 4 NVIDIA V100 GPUs of 16GB VRAM. Evaluation and results ---------------------- For evaluation details visit our GitHub repository. Citing ------ Check out our paper for all the details: URL
[]
[ "TAGS\n#transformers #pytorch #roberta #fill-mask #national library of spain #spanish #bne #es #dataset-bne #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ 83 ]
[ "passage: TAGS\n#transformers #pytorch #roberta #fill-mask #national library of spain #spanish #bne #es #dataset-bne #arxiv-1907.11692 #arxiv-2107.07253 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
[ -0.10817181318998337, 0.20310872793197632, -0.006971997208893299, 0.09333059936761856, 0.05503338202834129, 0.022737404331564903, 0.06438587605953217, 0.11308560520410538, 0.02791142277419567, 0.012531583197414875, 0.14112262427806854, 0.22381071746349335, 0.010571506805717945, 0.008289661258459091, -0.08142244815826416, -0.15719783306121826, 0.06974560767412186, 0.01848878152668476, -0.08204543590545654, 0.036870043724775314, 0.08250658214092255, -0.030994903296232224, 0.04517325013875961, -0.05160504952073097, -0.03761931508779526, 0.07400854676961899, 0.014425699599087238, -0.13304610550403595, 0.13990135490894318, 0.06758418679237366, 0.10124371200799942, 0.0445413701236248, -0.027157289907336235, -0.08710289746522903, 0.006310963071882725, -0.04196544736623764, -0.09918443113565445, 0.0781787857413292, 0.018764019012451172, -0.06481882929801941, 0.04297831282019615, 0.03447895124554634, -0.026929007843136787, 0.0074279047548770905, -0.16499505937099457, -0.22537167370319366, -0.10334708541631699, 0.033305153250694275, -0.0057276273146271706, 0.07397481054067612, 0.04957247152924538, 0.10527456551790237, -0.05688602477312088, 0.014960468746721745, 0.16726884245872498, -0.3073540925979614, -0.025846831500530243, -0.01473874319344759, 0.06603731960058212, 0.04028776288032532, 0.015542786568403244, 0.060197364538908005, 0.10865509510040283, -0.018139956519007683, -0.007054226938635111, -0.09804517030715942, -0.15076103806495667, 0.028438420966267586, -0.025940198451280594, -0.08537723124027252, 0.22159872949123383, -0.02874854765832424, 0.03485472872853279, 0.0580519363284111, -0.08136603981256485, 0.0910983458161354, 0.017499569803476334, 0.03539388254284859, 0.034378666430711746, 0.04033819958567619, 0.10044214129447937, -0.0031093547586351633, -0.08808539062738419, 0.020662104710936546, -0.20204627513885498, 0.11696501076221466, 0.0018451330251991749, 0.06272413581609726, -0.03950077295303345, 0.021890787407755852, -0.018595082685351372, -0.1384262591600418, 0.02123132534325123, -0.018524037674069405, 0.122124083340168, 0.06801588833332062, -0.04078782722353935, 0.014328009448945522, 0.11721952259540558, 0.2044961154460907, -0.039703261107206345, -0.032698024064302444, -0.04769380763173103, 0.12680770456790924, 0.013040932826697826, 0.041934024542570114, -0.015285451896488667, -0.07209529727697372, 0.06785821914672852, -0.09601213783025742, 0.07748092710971832, -0.0057055167853832245, -0.17865043878555298, -0.09782926738262177, -0.06658105552196503, 0.12312564998865128, 0.10291008651256561, -0.05367562919855118, -0.06393615156412125, 0.019723132252693176, 0.14587387442588806, -0.034242045134305954, 0.02972959168255329, -0.018756216391921043, -0.03434082865715027, 0.006088219117373228, -0.01841205172240734, -0.028426794335246086, -0.03085458092391491, 0.04406346380710602, -0.08017688244581223, -0.07496082037687302, -0.04340643063187599, -0.054637037217617035, 0.1115397959947586, -0.1217254102230072, 0.07803251594305038, -0.18560706079006195, -0.1579338163137436, 0.010052811354398727, 0.06558765470981598, -0.10760965198278427, -0.012998856604099274, -0.0015155710279941559, 0.02988150529563427, 0.0514078252017498, -0.07960808277130127, 0.022652389481663704, -0.09532810002565384, 0.10252062231302261, -0.005085853394120932, 0.0631510466337204, -0.19648045301437378, -0.0007645636796951294, -0.09707321226596832, 0.01741960644721985, -0.15052801370620728, -0.042944300919771194, -0.09413495659828186, 0.09573666751384735, -0.05946166068315506, -0.004398103803396225, -0.011833026073873043, 0.044620875269174576, 0.05559667572379112, 0.1107974424958229, -0.06946121156215668, -0.07565349340438843, 0.16395121812820435, -0.07688087224960327, -0.1625504046678543, 0.07853519171476364, 0.01990421861410141, 0.048818230628967285, 0.04024643078446388, 0.1737148016691208, -0.009688028134405613, -0.09789152443408966, 0.004176090471446514, 0.07328376919031143, -0.030238337814807892, -0.22561340034008026, 0.09054476022720337, -0.048603519797325134, -0.047133348882198334, 0.035051390528678894, -0.03128790482878685, 0.0635598674416542, -0.00667577376589179, -0.04690471291542053, 0.009684492833912373, -0.05871986970305443, 0.02816617861390114, 0.014038540422916412, 0.0671766921877861, -0.07428208738565445, 0.015555428341031075, -0.014663350768387318, 0.0001450292329536751, 0.07009927183389664, 0.04028483107686043, -0.027946189045906067, 0.1475709229707718, -0.039597898721694946, 0.0006510673556476831, -0.10217699408531189, 0.08661800622940063, -0.01876954734325409, 0.053557686507701874, -0.006174230016767979, 0.016953419893980026, 0.04180486872792244, -0.03926251828670502, -0.05770983546972275, -0.006780494004487991, 0.05839274451136589, 0.02437739446759224, 0.021552179008722305, -0.15763184428215027, 0.0817481279373169, -0.04363637790083885, -0.03930719941854477, -0.04905092343688011, -0.005098492372781038, -0.04902094230055809, 0.06223955750465393, -0.06248205155134201, 0.08961094915866852, -0.07487919926643372, 0.051991160959005356, -0.058300308883190155, -0.002397664589807391, 0.08782599121332169, 0.06030578538775444, -0.03352612629532814, 0.13904452323913574, -0.05833188816905022, 0.283641517162323, 0.16006268560886383, -0.1690659373998642, 0.023541074246168137, 0.007914029061794281, -0.0015134378336369991, 0.006697464268654585, 0.03769538551568985, -0.05052813142538071, 0.049736566841602325, -0.017058832570910454, 0.14538639783859253, -0.12199918925762177, -0.006456729490309954, 0.026665886864066124, -0.067608542740345, -0.06704895198345184, 0.11754345893859863, 0.1411515772342682, -0.12120527774095535, 0.18245482444763184, 0.23870179057121277, -0.05953988805413246, 0.15553289651870728, 0.0010357710998505354, -0.02619016356766224, 0.004115356598049402, -0.05562244728207588, 0.028425397351384163, 0.15218617022037506, -0.10601359605789185, 0.0370376780629158, 0.07006797939538956, -0.006439684424549341, 0.05681126192212105, -0.11265864223241806, -0.07563521713018417, 0.0034206300042569637, 0.0034477682784199715, -0.06157185137271881, 0.05143657699227333, -0.027348371222615242, 0.12242475897073746, 0.0214534904807806, -0.13486899435520172, 0.0690104216337204, 0.016006389632821083, -0.05202067643404007, 0.15921799838542938, -0.16179460287094116, -0.3060198426246643, -0.15914985537528992, -0.0773668885231018, 0.04522779583930969, 0.073092982172966, 0.0984559878706932, -0.043096475303173065, -0.02885107323527336, 0.06682335585355759, -0.017137788236141205, -0.03570364788174629, -0.06102203577756882, -0.01805918663740158, 0.08027146011590958, -0.03371107950806618, -0.13139154016971588, -0.029076414182782173, 0.04379578307271004, 0.011275232769548893, 0.035324886441230774, -0.10119086503982544, 0.1421440988779068, 0.03314097598195076, 0.04505542665719986, 0.010178578086197376, -0.028270365670323372, 0.11974744498729706, -0.04889913648366928, -0.014746599830687046, 0.16818737983703613, 0.02259882539510727, 0.02767498791217804, 0.16357144713401794, 0.038363054394721985, -0.05413886904716492, -0.0508139543235302, -0.05948392674326897, -0.0700814351439476, -0.2888064682483673, -0.11768869310617447, -0.08913568407297134, 0.04416080564260483, 0.06914616376161575, 0.05690670385956764, 0.11944372951984406, 0.08768186718225479, 0.023739423602819443, -0.00831670593470335, -0.0762610137462616, 0.06157086417078972, 0.22159914672374725, -0.012241634540259838, 0.06559343636035919, -0.08014220744371414, -0.049482520669698715, 0.12126415222883224, 0.12145759165287018, 0.05251067131757736, 0.08163748681545258, 0.09041855484247208, 0.024141816422343254, 0.14785702526569366, 0.009941987693309784, 0.10634104162454605, 0.025812938809394836, -0.029366400092840195, -0.0806649848818779, -0.02389652095735073, -0.059643570333719254, -0.013491760939359665, -0.01836290769279003, -0.03462062403559685, -0.05533825606107712, -0.22598662972450256, 0.058078814297914505, 0.045340459793806076, 0.046135421842336655, -0.1624487340450287, 0.020806951448321342, 0.058540452271699905, 0.022043200209736824, -0.10324778407812119, 0.0298799816519022, -0.03175424784421921, -0.11155913025140762, 0.06851822882890701, 0.025316720828413963, 0.12353329360485077, 0.009683790616691113, 0.03306538611650467, -0.1326194703578949, -0.11316040903329849, 0.03122185543179512, 0.1016756072640419, -0.24311187863349915, 0.31350943446159363, 0.023786695674061775, -0.0077596185728907585, -0.04317459464073181, -0.00841534323990345, -0.00013950227003078908, 0.10719068348407745, 0.12066899240016937, 0.02594013325870037, 0.013592536561191082, -0.009584920480847359, -0.03267190605401993, 0.04012689366936684, -0.08158920705318451, -0.02388579770922661, -0.04320641607046127, 0.002234712475910783, -0.01212928257882595, -0.01468789204955101, 0.08467079699039459, -0.000040834664105204865, -0.1532757431268692, 0.03312944993376732, 0.037237878888845444, 0.006582805421203375, -0.018114112317562103, -0.04424576088786125, -0.10179384052753448, 0.12925361096858978, -0.05290345475077629, -0.053517285734415054, -0.07407750189304352, -0.056111112236976624, 0.11954375356435776, -0.051811717450618744, 0.05964844673871994, -0.03253856301307678, -0.03517496585845947, -0.08935095369815826, -0.09134237468242645, 0.12569497525691986, -0.15165700018405914, 0.008372670970857143, -0.10362232476472855, 0.07550536096096039, -0.06283175945281982, 0.040086161345243454, 0.01594039797782898, 0.015317576937377453, -0.05658023804426193, -0.04569414258003235, 0.0241472776979208, -0.07941046357154846, 0.1317061334848404, -0.012372943572700024, -0.09904076904058456, -0.06987979263067245, 0.018766043707728386, -0.08652849495410919, 0.13333114981651306, 0.2562698423862457, -0.05168404057621956, 0.11424782872200012, 0.2733060419559479, -0.1169561818242073, -0.22521817684173584, -0.1369016319513321, -0.12080144137144089, -0.018031004816293716, -0.057368241250514984, -0.16948819160461426, 0.03436417877674103, 0.15529362857341766, -0.08302679657936096, 0.0850416049361229, -0.24666070938110352, -0.05932839587330818, 0.1503632664680481, -0.0030005150474607944, 0.41653740406036377, -0.11908061057329178, -0.09709174185991287, -0.10717098414897919, -0.1444958746433258, 0.1094001904129982, -0.05136117339134216, 0.037704531103372574, -0.03598437085747719, -0.04222622513771057, -0.009638707153499126, -0.021427446976304054, 0.1637425273656845, -0.08329187333583832, -0.010542508214712143, -0.06988102942705154, -0.03455633297562599, 0.12281136214733124, -0.007728678174316883, -0.017965659499168396, -0.06059487536549568, -0.004578818567097187, -0.13814301788806915, -0.0037811617366969585, -0.07760240882635117, 0.08373400568962097, -0.019743552431464195, -0.02651406079530716, -0.023969998583197594, -0.0009343305719085038, 0.0057798526249825954, -0.006028153467923403, 0.16817377507686615, 0.033949077129364014, 0.09371142834424973, 0.14244361221790314, 0.04760768637061119, -0.09897590428590775, 0.010091078467667103, -0.040613092482089996, -0.06316415965557098, 0.07718313485383987, -0.036711275577545166, 0.011431853286921978, 0.12247196584939957, -0.04685593023896217, 0.024640360847115517, 0.04018067196011543, -0.033217839896678925, -0.0193686094135046, 0.15400540828704834, -0.10077078640460968, 0.07469917833805084, 0.014176069758832455, 0.04720723628997803, 0.08933867514133453, -0.017212139442563057, 0.10418465733528137, 0.04096699506044388, -0.0545668825507164, 0.01672716811299324, -0.024263354018330574, -0.04546774551272392, 0.09527066349983215, 0.06885433942079544, -0.00006753960042260587, -0.10021428763866425, 0.09361714124679565, -0.00035994272911921144, -0.18267902731895447, 0.005553741008043289, 0.09958399832248688, -0.05338464304804802, -0.09836888313293457, 0.044450175017118454, 0.09538974612951279, -0.2777233123779297, -0.12030315399169922, -0.14167824387550354, -0.07271117717027664, 0.06314882636070251, 0.21333011984825134, 0.05926893651485443, 0.015700865536928177, 0.021206650882959366, -0.027455834671854973, 0.04960489273071289, -0.009198910556733608, -0.05697852745652199, 0.024983102455735207, -0.04658493399620056, -0.07104982435703278, -0.003042182419449091, 0.028430908918380737, -0.04319439083337784, 0.04843815043568611, -0.1761520653963089, 0.046321313828229904, -0.13682974874973297, 0.04149448499083519, -0.08104518055915833, -0.044697873294353485, -0.0419221892952919, -0.08323553204536438, -0.03899591043591499, -0.06855668127536774, -0.0966012254357338, 0.004618645645678043, 0.02510862797498703, 0.0683593899011612, -0.08092878013849258, -0.054927758872509, 0.07724858075380325, -0.011857801117002964, 0.05876767635345459, 0.07942426204681396, -0.002903212094679475, 0.10072943568229675, -0.1843300312757492, -0.04844437912106514, 0.07801813632249832, 0.049724314361810684, 0.05745372548699379, -0.0032924681436270475, 0.011291739530861378, 0.10495751351118088, -0.015450369566679, 0.0484817810356617, -0.02553027868270874, -0.1281830072402954, 0.0280994214117527, 0.056890591979026794, -0.1978885978460312, 0.03488048538565636, -0.021329576149582863, 0.1822541058063507, -0.05559267848730087, 0.07007995992898941, -0.060960303992033005, -0.0026777959428727627, -0.03146453946828842, 0.028087852522730827, -0.01897263340651989, -0.09729664027690887, -0.08406274765729904, -0.02144882082939148, -0.022760894149541855, 0.012463339604437351, 0.24773356318473816, 0.02878081612288952, -0.022428365424275398, 0.027011040598154068, 0.015260271728038788, -0.010643012821674347, 0.013154328800737858, 0.23743028938770294, 0.06911129504442215, -0.01155712641775608, -0.139927476644516, 0.08704622834920883, 0.03095470555126667, 0.0397803857922554, 0.026455407962203026, 0.1462932825088501, 0.2702261805534363, 0.07365474849939346, 0.05616656690835953, -0.028700444847345352, 0.0015965357888489962, -0.08856857568025589, 0.0730258971452713, 0.030638987198472023, 0.07038411498069763, 0.07882748544216156, 0.16647955775260925, -0.04575353488326073, 0.038287799805402756, -0.030874252319335938, 0.022410571575164795, -0.11511614173650742, -0.09438202530145645, -0.03993077576160431, -0.0910993367433548, -0.01092811580747366, -0.06014535203576088, 0.04263549670577049, 0.024602459743618965, 0.023475442081689835, -0.06346417963504791, -0.07940821349620819, 0.06104256212711334, -0.08281167596578598, 0.01737947016954422, 0.03951617330312729, 0.09022513031959534, -0.09439235180616379, 0.022808320820331573, -0.0862647294998169, 0.058885328471660614, -0.0434952974319458, 0.05588072910904884, 0.003622537013143301, 0.009331348352134228, -0.05551844462752342, -0.03616322949528694, -0.030741186812520027, 0.05308174341917038, 0.017511676996946335, 0.1843118518590927, -0.0061880419962108135, 0.039190638810396194, 0.06121756508946419, 0.20344707369804382, -0.022450828924775124, -0.06977364420890808, -0.04378997161984444, 0.07879894971847534, 0.015052388422191143, 0.0980411171913147, -0.013988901861011982, 0.0012219490017741919, -0.05717403069138527, 0.2495289444923401, 0.2928844392299652, -0.0397014245390892, -0.008310964331030846, 0.04620962589979172, 0.02411886677145958, 0.08754832297563553, 0.05432867258787155, 0.09651612490415573, 0.2969628572463989, -0.0976187065243721, -0.07883062958717346, -0.09446041285991669, 0.07065851241350174, -0.12698574364185333, 0.09976290911436081, -0.042604461312294006, -0.09925293177366257, -0.009981575421988964, 0.10233666747808456, -0.09637825191020966, -0.04688365012407303, 0.0474868044257164, -0.19230656325817108, -0.09537660330533981, -0.022195814177393913, 0.13683685660362244, 0.05986500903964043, 0.0628414899110794, -0.02302134968340397, -0.05936474725604057, 0.0068147312849760056, 0.0038668832276016474, -0.1779661327600479, -0.15910528600215912, 0.04968655854463577, 0.01929519511759281, 0.2247513234615326, -0.025113116949796677, 0.025674039497971535, 0.11185403913259506, 0.05785488337278366, -0.10148259997367859, 0.03135709837079048, 0.06966892629861832, -0.034127138555049896, 0.05687996372580528, -0.1347975879907608, 0.010469846427440643, -0.031402587890625, 0.05596665292978287, -0.053629446774721146, 0.05869147926568985, 0.10633812099695206, -0.04445892944931984, -0.049189966171979904, 0.11179493367671967, -0.10284868627786636, 0.02670038677752018, 0.0482979491353035, -0.0069682784378528595, -0.04284790903329849, -0.032642826437950134, -0.008706665597856045, 0.08358263969421387, -0.03475720435380936, -0.0770348310470581, -0.07667385786771774, -0.02352266199886799, 0.05469944700598717, 0.000598185695707798, -0.15249764919281006, 0.019031433388590813, -0.10293423384428024, -0.00837673433125019, -0.1356157511472702, -0.011688479222357273, 0.006212752312421799, 0.018764175474643707, 0.012203912250697613, -0.0772874727845192, -0.00635465607047081, 0.005130188073962927, -0.04268669709563255, -0.06864291429519653 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-base-timit-demo-colab This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4877 - Wer: 0.4895 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 3.6615 | 4.0 | 500 | 1.7423 | 1.0723 | | 0.8519 | 8.0 | 1000 | 0.4877 | 0.4895 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "wav2vec2-base-timit-demo-colab", "results": []}]}
automatic-speech-recognition
BSen/wav2vec2-base-timit-demo-colab
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us
wav2vec2-base-timit-demo-colab ============================== This model is a fine-tuned version of facebook/wav2vec2-base on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.4877 * Wer: 0.4895 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0001 * train\_batch\_size: 32 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 1000 * num\_epochs: 10 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.10.0+cu111 * Datasets 1.13.3 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.13.3\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.13.3\n* Tokenizers 0.10.3" ]
[ 56, 130, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.13.3\n* Tokenizers 0.10.3" ]
[ -0.10845157504081726, 0.09908455610275269, -0.003328087506815791, 0.064132459461689, 0.10895244032144547, -0.019833922386169434, 0.12862396240234375, 0.15069493651390076, -0.09068844467401505, 0.07495837658643723, 0.12558989226818085, 0.15130189061164856, 0.04247428849339485, 0.14587485790252686, -0.04989681392908096, -0.2831854522228241, 0.047148410230875015, 0.0344185009598732, -0.011495736427605152, 0.12580420076847076, 0.08374413102865219, -0.1246558427810669, 0.0575408972799778, 0.03439861163496971, -0.15762007236480713, -0.004222361836582422, -0.005132773891091347, -0.10503344237804413, 0.12318093329668045, 0.005175852682441473, 0.07036325335502625, 0.04895547777414322, 0.06533050537109375, -0.21929295361042023, 0.00653934245929122, 0.043978992849588394, 0.028142070397734642, 0.07395869493484497, 0.05927210673689842, -0.02922554686665535, 0.10316165536642075, -0.07494337111711502, 0.08143699169158936, 0.03787862882018089, -0.10527483373880386, -0.2927303612232208, -0.08582467585802078, 0.048020146787166595, 0.06906755268573761, 0.08773531019687653, -0.012603127397596836, 0.14334647357463837, -0.05513891577720642, 0.11018762737512589, 0.2818458080291748, -0.3132283389568329, -0.04464263468980789, -0.03912581875920296, 0.05801953375339508, 0.06017890200018883, -0.10060273110866547, -0.018389474600553513, 0.015309194102883339, 0.04469738528132439, 0.13822920620441437, -0.016099687665700912, -0.059912946075201035, -0.007271978538483381, -0.14898927509784698, -0.059371236711740494, 0.11519980430603027, 0.02237517200410366, -0.03902164474129677, -0.09970881789922714, -0.05519965663552284, -0.2135244458913803, -0.06773342937231064, -0.016692254692316055, 0.04259074479341507, -0.04309910908341408, -0.10335616022348404, -0.01125268917530775, -0.0668254941701889, -0.07371607422828674, -0.03895795717835426, 0.18903189897537231, 0.05753960460424423, -0.00218514958396554, -0.03802049160003662, 0.07617509365081787, -0.021437905728816986, -0.138326957821846, -0.02440088614821434, 0.03546912223100662, -0.021839234977960587, -0.015550471842288971, -0.041658420115709305, -0.05788406357169151, 0.022332163527607918, 0.1626707911491394, -0.09934304654598236, 0.09706015139818192, -0.02053230255842209, 0.039010029286146164, -0.10296079516410828, 0.207469180226326, -0.04154599830508232, 0.017830070108175278, -0.009430598467588425, 0.05654627084732056, 0.029343798756599426, -0.02637169323861599, -0.094449482858181, 0.031710587441921234, 0.12202340364456177, 0.047265201807022095, -0.04700891673564911, 0.06585006415843964, -0.03330715745687485, -0.010179128497838974, 0.001268826425075531, -0.1126072034239769, 0.03694586455821991, 0.02089841663837433, -0.06531715393066406, 0.003966223448514938, 0.014248437248170376, 0.007738653104752302, -0.054581139236688614, 0.08320555835962296, -0.061709631234407425, 0.033214226365089417, -0.0567132942378521, -0.12704133987426758, 0.026138080283999443, -0.11777470260858536, -0.0032138400711119175, -0.1004684567451477, -0.10051196813583374, -0.012233131565153599, 0.03638278692960739, -0.03764696791768074, -0.025255704298615456, -0.07906321436166763, -0.09166959673166275, 0.04490082710981369, -0.033869579434394836, 0.07094687968492508, -0.07456310838460922, 0.09287424385547638, 0.033206965774297714, 0.0878080427646637, -0.014713220298290253, 0.06028376519680023, -0.07103770226240158, 0.02708076685667038, -0.19965869188308716, 0.07535413652658463, -0.08978839218616486, 0.0594932921230793, -0.12448719143867493, -0.11428660154342651, 0.0228386539965868, -0.007102643139660358, 0.09809787571430206, 0.09807764738798141, -0.17203669250011444, -0.08839897811412811, 0.20844553411006927, -0.08258125185966492, -0.0839410051703453, 0.12442778795957565, -0.025004802271723747, -0.0002877268416341394, 0.05687669664621353, 0.2589298486709595, 0.04792969301342964, -0.12577976286411285, 0.006772120948880911, -0.04189281165599823, 0.04327783361077309, -0.03484630212187767, 0.0582137256860733, -0.0272817425429821, 0.06823428720235825, 0.01882919669151306, -0.003709828481078148, 0.03688151389360428, -0.08667941391468048, -0.07698243856430054, -0.044127047061920166, -0.07779230177402496, 0.030667150393128395, 0.033142685890197754, 0.06355591863393784, -0.11766107380390167, -0.10817836225032806, 0.03837360814213753, 0.0809110626578331, -0.10460010915994644, 0.07127842307090759, -0.11981219798326492, 0.0844627320766449, -0.014534168876707554, -0.004781613126397133, -0.18978941440582275, 0.035016387701034546, 0.038745298981666565, -0.030518589541316032, 0.03949175029993057, -0.06401310116052628, 0.07811789959669113, 0.04584262892603874, -0.02613617293536663, -0.04677101969718933, -0.008521582931280136, 0.01106437761336565, -0.09017810225486755, -0.20698976516723633, -0.038269974291324615, -0.038506072014570236, 0.07914508879184723, -0.14031587541103363, 0.034840766340494156, 0.07770416885614395, 0.09242900460958481, 0.03307691216468811, -0.03158583864569664, -0.0004921602667309344, 0.08935478329658508, -0.02080194652080536, -0.06457395851612091, 0.057812537997961044, 0.018930960446596146, -0.0864320695400238, 0.03844248875975609, -0.1512802392244339, 0.12604403495788574, 0.14709104597568512, -0.013463634066283703, -0.06609942764043808, 0.00041650334605947137, -0.04794585332274437, -0.03473477065563202, -0.002966534346342087, 0.03220808878540993, 0.21481148898601532, 0.013312424533069134, 0.14296689629554749, -0.08898089826107025, -0.04231981933116913, 0.050421297550201416, -0.020980870351195335, -0.0067395796068012714, 0.11755865067243576, 0.045584868639707565, -0.0530962236225605, 0.11892364174127579, 0.09137091040611267, -0.07877188175916672, 0.1202922835946083, -0.06040457263588905, -0.07432533800601959, -0.019718555733561516, 0.005360029637813568, 0.023622620850801468, 0.0979253277182579, -0.1644127517938614, -0.040304336696863174, 0.026283739134669304, 0.025855904445052147, 0.020282864570617676, -0.20866768062114716, 0.014641070738434792, 0.028371281921863556, -0.08456819504499435, -0.04375169798731804, 0.0024561879690736532, 0.01297522522509098, 0.09419545531272888, 0.011378164403140545, -0.09415283054113388, 0.010385503061115742, 0.004013699945062399, -0.07340918481349945, 0.17607977986335754, -0.11754605919122696, -0.17664648592472076, -0.10378023236989975, -0.09281601756811142, -0.03961852565407753, -0.002668161876499653, 0.08980774879455566, -0.09224725514650345, -0.03969765082001686, -0.0841042771935463, -0.0160831268876791, -0.02616148255765438, 0.04184986278414726, 0.030495068058371544, -0.010968036018311977, 0.06309769302606583, -0.11713184416294098, -0.021555928513407707, -0.03967585787177086, -0.00165739085059613, 0.05473687872290611, 0.037354134023189545, 0.10816796123981476, 0.15858960151672363, -0.009759564884006977, 0.050042763352394104, -0.04675820842385292, 0.18646715581417084, -0.07504145801067352, -0.0361141636967659, 0.10935748368501663, -0.006430997978895903, 0.06889061629772186, 0.11922606825828552, 0.047692738473415375, -0.09753765165805817, -0.013566586188971996, 0.003516223281621933, -0.04507129639387131, -0.21587467193603516, -0.03453608602285385, -0.044885143637657166, 0.0005507901078090072, 0.10616638511419296, 0.04133860394358635, 0.039367299526929855, 0.022594867274165154, 0.031980086117982864, 0.005526726599782705, 0.004311400465667248, 0.09672193229198456, 0.12982311844825745, 0.03950730711221695, 0.1332462877035141, -0.03838793933391571, -0.03653016686439514, 0.029464080929756165, 0.0051406379789114, 0.2320398986339569, 0.019559642300009727, 0.19054539501667023, 0.0568314753472805, 0.17481017112731934, 0.041123166680336, 0.06727857887744904, -0.0009883451275527477, -0.01100933738052845, 0.011673470959067345, -0.05223064124584198, -0.03824496269226074, 0.024140696972608566, 0.02445647120475769, 0.010015109553933144, -0.11442259699106216, -0.012498073279857635, 0.047190431505441666, 0.3504691421985626, 0.028892485424876213, -0.3376957178115845, -0.09010003507137299, -0.0117866275832057, -0.08601811528205872, -0.029619377106428146, 0.045725882053375244, 0.08935139328241348, -0.08077379316091537, 0.06510791182518005, -0.06270785629749298, 0.08982740342617035, -0.0632234439253807, 0.03450314700603485, 0.038002099841833115, 0.07199987769126892, 0.00403654994443059, 0.03387007117271423, -0.29303649067878723, 0.27988818287849426, 0.005018804222345352, 0.0770687460899353, -0.061891257762908936, 0.00849767867475748, 0.025893045589327812, 0.01698690839111805, 0.08719101548194885, -0.026376822963356972, -0.12230709195137024, -0.1753503382205963, -0.0930066704750061, 0.012037490494549274, 0.12901856005191803, 0.012934127822518349, 0.11074932664632797, -0.010881232097744942, -0.017629779875278473, 0.0495753176510334, -0.09598478674888611, -0.06587833911180496, -0.09221106767654419, 0.012702595442533493, 0.0815073549747467, 0.033197157084941864, -0.07232941687107086, -0.10335580259561539, -0.09066176414489746, 0.14836853742599487, -0.053927887231111526, -0.04300789162516594, -0.11874590069055557, 0.007628877647221088, 0.11033546924591064, -0.07941905409097672, 0.06216587498784065, 0.009646382182836533, 0.10390357673168182, 0.011267532594501972, -0.0686769112944603, 0.12009028345346451, -0.06438003480434418, -0.16830359399318695, -0.02865135669708252, 0.14618822932243347, 0.029516080394387245, 0.06014872342348099, -0.00724834343418479, 0.03836947679519653, -0.021320756524801254, -0.07724367082118988, 0.04057418555021286, 0.028738563880324364, 0.04403327777981758, -0.013946851715445518, -0.020091257989406586, -0.008319968357682228, -0.09151263535022736, -0.0172145813703537, 0.20653504133224487, 0.24275268614292145, -0.09658938646316528, 0.09374846518039703, 0.0687519907951355, -0.04289892315864563, -0.17178988456726074, -0.003977533895522356, 0.06532268226146698, -0.0005410643643699586, -0.024695495143532753, -0.19317547976970673, 0.02393139898777008, 0.0702742412686348, -0.020368129014968872, 0.08475038409233093, -0.3191103935241699, -0.1406862437725067, 0.13666577637195587, 0.11406727880239487, 0.058078426867723465, -0.14678438007831573, -0.055577460676431656, -0.0098577830940485, -0.10320964455604553, 0.09486517310142517, -0.07534069567918777, 0.13563251495361328, -0.02317049726843834, 0.08875183016061783, 0.011715149506926537, -0.05837837606668472, 0.10626716911792755, 0.013421930372714996, 0.06000230461359024, -0.04604509845376015, 0.016785942018032074, 0.04842709004878998, -0.06310348212718964, 0.05501803383231163, -0.07990279793739319, 0.027067571878433228, -0.07999729365110397, -0.03261750936508179, -0.08482589572668076, 0.014049052260816097, -0.009570451453328133, -0.03293005749583244, -0.03662450984120369, 0.00107376289088279, 0.063186414539814, -0.010509365238249302, 0.15528656542301178, -0.0270612183958292, 0.12730875611305237, 0.16075323522090912, 0.10111255943775177, -0.103737011551857, -0.07711600512266159, 0.005530905909836292, -0.03423431143164635, 0.054409466683864594, -0.11796093732118607, 0.03801124542951584, 0.1359485685825348, 0.031251613050699234, 0.12257102876901627, 0.06995140761137009, -0.06538712233304977, 0.0338173434138298, 0.041817259043455124, -0.13711358606815338, -0.12800286710262299, 0.013914138078689575, 0.022808127105236053, -0.07214707881212234, 0.07436768710613251, 0.11520446836948395, -0.05602113530039787, -0.014180196449160576, -0.002624042797833681, 0.014660888351500034, -0.03975299000740051, 0.19580353796482086, 0.036799702793359756, 0.061699967831373215, -0.12463726103305817, 0.08149527758359909, 0.03910529613494873, -0.13564395904541016, 0.06075034663081169, 0.10620077699422836, -0.09526684880256653, -0.029285304248332977, 0.02825722098350525, 0.11132549494504929, -0.0238035898655653, -0.07368021458387375, -0.14215122163295746, -0.14393025636672974, 0.10869693011045456, 0.20556962490081787, 0.05594154819846153, 0.017187677323818207, -0.05789783596992493, 0.017771124839782715, -0.11888305842876434, 0.07013432681560516, 0.0412360243499279, 0.060645852237939835, -0.12966282665729523, 0.1466074436903, 0.017476215958595276, 0.03952096030116081, -0.014505460858345032, -0.011100616306066513, -0.11256121844053268, 0.038844965398311615, -0.1278262883424759, 0.005290407687425613, -0.06745150685310364, 0.0008465367136523128, 0.004387800581753254, -0.049631666392087936, -0.06365000456571579, 0.03404141589999199, -0.12005818635225296, -0.02337501384317875, 0.0010936331236734986, 0.037271320819854736, -0.12881337106227875, -0.010186615400016308, 0.01422447245568037, -0.09466855973005295, 0.09787091612815857, 0.08696584403514862, -0.03332767263054848, 0.05084272846579552, -0.061657316982746124, -0.026208769530057907, 0.07925371080636978, -0.006773363333195448, 0.050752412527799606, -0.13069352507591248, -0.019264893606305122, 0.010627862066030502, 0.03486645221710205, 0.02407766506075859, 0.11132092773914337, -0.11559705436229706, 0.00031990036950446665, -0.02796434983611107, -0.052222996950149536, -0.06826756149530411, 0.049667950719594955, 0.10854289680719376, 0.025416718795895576, 0.16421833634376526, -0.09300988167524338, 0.02963634766638279, -0.1661534458398819, 0.005690810736268759, -0.015712391585111618, -0.12261787056922913, -0.04941120743751526, -0.03158865123987198, 0.07850610464811325, -0.06370696425437927, 0.12990696728229523, -0.0313095860183239, 0.02621540240943432, 0.03710032254457474, -0.07606440782546997, -0.053592290729284286, 0.041270799934864044, 0.20473076403141022, 0.038485087454319, -0.04331443831324577, 0.07334163039922714, 0.02112748473882675, 0.08146321773529053, 0.12666654586791992, 0.17341424524784088, 0.16052812337875366, 0.06224586069583893, 0.11664967238903046, 0.05435524135828018, -0.05371146276593208, -0.17385505139827728, 0.09297295659780502, -0.060667864978313446, 0.13134072721004486, -0.014764891937375069, 0.24152272939682007, 0.1204805076122284, -0.15356649458408356, 0.0658462792634964, -0.01941412128508091, -0.08950360864400864, -0.11543282121419907, -0.06576383113861084, -0.086906798183918, -0.1757485419511795, 0.008838088251650333, -0.10177844762802124, 0.061986472457647324, 0.0474947988986969, 0.03780747577548027, 0.016737710684537888, 0.13787272572517395, 0.015500339679419994, 0.003068692982196808, 0.09179627150297165, -0.002924825996160507, -0.05678891763091087, -0.0727107897400856, -0.08402785658836365, 0.034350812435150146, -0.013115049339830875, 0.05793365463614464, -0.004492188338190317, -0.0693507120013237, 0.04785351827740669, -0.039620932191610336, -0.09664297103881836, 0.023551076650619507, 0.021050360053777695, 0.06944486498832703, 0.049895357340574265, 0.034841954708099365, -0.04253607615828514, -0.0019231364130973816, 0.1950521618127823, -0.09436342120170593, -0.09340988844633102, -0.10991383343935013, 0.2521773874759674, 0.03970712423324585, -0.01640821062028408, 0.022032061591744423, -0.060897089540958405, -0.03063531592488289, 0.2122863382101059, 0.17326310276985168, -0.008460547775030136, 0.004555781837552786, -0.014195537194609642, -0.006267039105296135, -0.03690578043460846, 0.07991893589496613, 0.14700472354888916, 0.061061762273311615, -0.0627540647983551, -0.05133674666285515, -0.04929935559630394, -0.03518056124448776, -0.06757284700870514, 0.07585172355175018, 0.006315961945801973, -0.025491269305348396, -0.04407088831067085, 0.06440035998821259, -0.09428298473358154, -0.08201784640550613, 0.024687916040420532, -0.19545882940292358, -0.14933113753795624, 0.0055741737596690655, 0.07196994125843048, 0.012468140572309494, 0.03491340950131416, 0.003406234085559845, -0.010522718541324139, 0.0818401500582695, -0.0013626479776576161, -0.08105544000864029, -0.06557899713516235, 0.08365567028522491, -0.13297702372074127, 0.16666866838932037, -0.04247690364718437, 0.04812336713075638, 0.12315264344215393, 0.08853962272405624, -0.08068468421697617, 0.08553867787122726, 0.04326828941702843, -0.10737127810716629, 0.021810045465826988, 0.15285494923591614, -0.03375370055437088, 0.09370043128728867, 0.030894815921783447, -0.1144028976559639, 0.015156601555645466, -0.09005659818649292, -0.0380605086684227, -0.04064904525876045, -0.04891986399888992, -0.04352889582514763, 0.10911062359809875, 0.16310331225395203, -0.04403587058186531, 0.0038610929623246193, -0.05250238999724388, 0.011125252582132816, 0.04663541540503502, -0.0009615443996153772, -0.061863187700510025, -0.2793128490447998, 0.01147431880235672, 0.03637845441699028, 0.003821633756160736, -0.255357950925827, -0.09666264802217484, 0.012994683347642422, -0.043355971574783325, -0.08905532211065292, 0.08536598086357117, 0.07533194869756699, 0.04598110914230347, -0.05274122208356857, -0.058737147599458694, -0.0354306735098362, 0.18995919823646545, -0.17525547742843628, -0.05977578088641167 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-large-xls-r-300m-turkish-colab This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "wav2vec2-large-xls-r-300m-turkish-colab", "results": []}]}
automatic-speech-recognition
BSen/wav2vec2-large-xls-r-300m-turkish-colab
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "dataset:common_voice", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us
# wav2vec2-large-xls-r-300m-turkish-colab This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the common_voice dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
[ "# wav2vec2-large-xls-r-300m-turkish-colab\n\nThis model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the common_voice dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0003\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- num_epochs: 30\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n", "# wav2vec2-large-xls-r-300m-turkish-colab\n\nThis model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the common_voice dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0003\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- num_epochs: 30\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ 65, 53, 6, 12, 8, 3, 140, 4, 33 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n# wav2vec2-large-xls-r-300m-turkish-colab\n\nThis model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the common_voice dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0003\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- gradient_accumulation_steps: 2\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 500\n- num_epochs: 30\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ -0.08907194435596466, 0.1308734118938446, -0.002234972082078457, 0.02376195415854454, 0.13202445209026337, 0.02365260384976864, 0.10974742472171783, 0.11843730509281158, -0.09215465933084488, 0.09087226539850235, 0.06767266988754272, 0.0004319735162425786, 0.10014966875314713, 0.08499716967344284, 0.013344360515475273, -0.25029489398002625, -0.0014033388579264283, -0.019197087734937668, -0.05514322221279144, 0.09491205960512161, 0.11522868275642395, -0.07009121030569077, 0.0236611720174551, 0.025804826989769936, -0.1318603903055191, 0.03782056272029877, -0.08170772343873978, -0.0796193778514862, 0.08529306203126907, 0.02079794742166996, 0.048674389719963074, -0.00943757127970457, 0.08638788759708405, -0.26995861530303955, 0.0036886923480778933, 0.05278274789452553, 0.02670833095908165, 0.05500909313559532, 0.0945362076163292, -0.006062474567443132, 0.14133483171463013, -0.18938124179840088, 0.07696539908647537, 0.024544881656765938, -0.04347716644406319, -0.17004571855068207, -0.09653598070144653, 0.10879618674516678, 0.1270260512828827, 0.11959144473075867, -0.02248121052980423, 0.10708767920732498, -0.07384903728961945, 0.0690515860915184, 0.16194511950016022, -0.27587565779685974, -0.06929660588502884, -0.024262961000204086, 0.037173446267843246, 0.06702403724193573, -0.08925081789493561, 0.01023961789906025, 0.030231419950723648, 0.013848381116986275, 0.027510209009051323, 0.013237718492746353, -0.022487051784992218, -0.026729822158813477, -0.10388030856847763, -0.04294656962156296, 0.19394579529762268, 0.09995023161172867, -0.025019599124789238, -0.16427317261695862, 0.019751738756895065, -0.1185300424695015, -0.030053913593292236, -0.02168898656964302, -0.0001289581850869581, -0.044335201382637024, -0.0832374319434166, -0.01454573031514883, -0.06970511376857758, -0.04022286832332611, 0.04328522831201553, 0.1285329908132553, 0.04393918439745903, -0.02652803622186184, 0.004881024360656738, 0.07677800953388214, 0.03528212383389473, -0.12239225208759308, -0.001774440286681056, 0.041356008499860764, -0.1325727254152298, -0.04690669849514961, -0.033686671406030655, -0.08739448338747025, 0.005546235479414463, 0.09449991583824158, 0.035737328231334686, 0.08483070880174637, 0.02214752696454525, 0.002195083536207676, -0.01685996539890766, 0.132343128323555, -0.04582276940345764, -0.09757553786039352, -0.04938381910324097, 0.07849527895450592, 0.00480051152408123, -0.011865932494401932, -0.07696793228387833, -0.022960416972637177, 0.10459132492542267, 0.08559742569923401, -0.04146288335323334, -0.006012352183461189, -0.02810376137495041, -0.023889783769845963, 0.01423230767250061, -0.12370411306619644, 0.04487494006752968, -0.015085008926689625, -0.056830476969480515, 0.007386064622551203, -0.032452065497636795, 0.014972327277064323, -0.05248742178082466, 0.08590900897979736, -0.038751270622015, -0.011018029414117336, -0.03603692725300789, -0.03346796706318855, 0.03798522427678108, -0.028035899624228477, 0.012479260563850403, -0.07120301574468613, -0.11444729566574097, -0.048031508922576904, 0.03433554247021675, -0.07072102278470993, -0.06079639121890068, -0.029871944338083267, -0.017700904980301857, 0.033988796174526215, -0.015248429030179977, 0.1597096472978592, -0.041900575160980225, 0.06956243515014648, -0.013590122573077679, 0.009052556939423084, 0.08338887989521027, 0.05796118453145027, -0.05249861255288124, 0.030015945434570312, -0.03716098144650459, 0.10235199332237244, -0.08292912691831589, 0.033611029386520386, -0.14574456214904785, -0.09583645313978195, -0.04093651473522186, -0.03398391976952553, 0.0670052021741867, 0.0877293273806572, -0.1641279011964798, -0.057152558118104935, 0.15633606910705566, -0.0439581461250782, -0.09313443303108215, 0.1313215047121048, -0.01241466123610735, -0.013778147287666798, 0.07156288623809814, 0.1498272716999054, 0.12407780438661575, -0.09636641293764114, -0.0652654841542244, -0.020026346668601036, 0.09068748354911804, 0.03745729476213455, 0.09930124133825302, -0.030399218201637268, 0.041977282613515854, 0.010576204396784306, -0.012726499699056149, 0.036864280700683594, -0.0536394827067852, -0.08648020029067993, -0.015263252891600132, -0.09184212982654572, 0.008136502467095852, 0.022478321567177773, 0.03396435081958771, -0.10202289372682571, -0.14340604841709137, 0.05726740136742592, 0.14016897976398468, -0.05544974282383919, 0.0019858432933688164, -0.0728345587849617, 0.013912774622440338, -0.05749351158738136, -0.012877849861979485, -0.16813117265701294, -0.048593539744615555, 0.037639521062374115, -0.0814872533082962, 0.05089573562145233, -0.013952745124697685, 0.056913264095783234, 0.0407119020819664, -0.04671680927276611, -0.020242925733327866, -0.09246945381164551, 0.00985008291900158, -0.06817512959241867, -0.13773763179779053, -0.053359221667051315, -0.023775996640324593, 0.23491156101226807, -0.18798376619815826, -0.003677451517432928, 0.03172088786959648, 0.15431629121303558, -0.0016732985386624932, -0.071092389523983, 0.0039141615852713585, 0.051379092037677765, 0.0202434454113245, -0.09018826484680176, 0.007623758167028427, 0.0016014075372368097, -0.1263660192489624, -0.04972507804632187, -0.13163122534751892, 0.03984267637133598, 0.07747333496809006, 0.11805378645658493, -0.0923546776175499, -0.06411784887313843, -0.05868242681026459, -0.052662041038274765, -0.04718293994665146, -0.0181781854480505, 0.23247896134853363, 0.03931906074285507, 0.09123118966817856, -0.04441050812602043, -0.06838392466306686, 0.017566274851560593, 0.029673755168914795, -0.05736365541815758, 0.09102202951908112, 0.02494465373456478, -0.1546468734741211, 0.07349096238613129, 0.06876899302005768, -0.014502009376883507, 0.15265078842639923, -0.05156100168824196, -0.1129584014415741, -0.031888823956251144, 0.022456718608736992, 0.0130989458411932, 0.10543818771839142, -0.14352168142795563, -0.01564243994653225, 0.036594852805137634, 0.004730363376438618, 0.04059974104166031, -0.12406180799007416, -0.004022275097668171, 0.0513615719974041, -0.027397818863391876, -0.015613551251590252, -0.022294722497463226, 0.0049219592474401, 0.06590204685926437, 0.033718522638082504, 0.01988946460187435, 0.005998822394758463, -0.01918071322143078, -0.0832522064447403, 0.13603529334068298, -0.09992313385009766, -0.19224536418914795, -0.13219423592090607, 0.04046912491321564, -0.04613475501537323, -0.05018068850040436, 0.026212187483906746, -0.12920789420604706, -0.04255163297057152, -0.061849478632211685, -0.012819874100387096, -0.07914286106824875, 0.0036393951158970594, 0.06776531785726547, 0.018147306516766548, 0.0711691826581955, -0.11059547960758209, 0.027318870648741722, 0.006860899738967419, -0.04467171058058739, -0.03156488761305809, 0.022933967411518097, 0.10358745604753494, 0.11424480378627777, -0.012410471215844154, 0.034129075706005096, -0.028570910915732384, 0.1697940081357956, -0.12666288018226624, -0.01342800259590149, 0.09597828984260559, 0.01815725304186344, 0.034074701368808746, 0.09631989151239395, 0.033172667026519775, -0.07034839689731598, 0.019703073427081108, 0.05374772846698761, -0.022251121699810028, -0.2407681941986084, -0.062301844358444214, -0.05599098280072212, -0.11524193733930588, 0.1219988688826561, 0.06393402814865112, -0.01540953479707241, 0.05000283196568489, -0.04416794329881668, 0.034638915210962296, 0.0021611510310322046, 0.06991741061210632, 0.06541234999895096, 0.05855443328619003, 0.07271253317594528, -0.04538705199956894, -0.039334289729595184, 0.05381649732589722, 0.021625405177474022, 0.24025483429431915, 0.01636429876089096, 0.15692384541034698, 0.01302530150860548, 0.12511269748210907, -0.01608767919242382, 0.024631451815366745, 0.028433647006750107, -0.009206329472362995, 0.03330882638692856, -0.06954026967287064, -0.02646121196448803, 0.033907145261764526, 0.07979468256235123, 0.0070115611888468266, -0.07944858074188232, 0.01569763943552971, 0.03088078275322914, 0.25573158264160156, 0.04458342120051384, -0.2651447355747223, -0.0837121531367302, 0.005899876356124878, -0.06493585556745529, -0.06091846525669098, -0.0018304719123989344, 0.08173611760139465, -0.13624487817287445, 0.0985512062907219, -0.04862414300441742, 0.09598607569932938, -0.0685688778758049, -0.014315848238766193, 0.05122587829828262, 0.07156499475240707, 0.008973106741905212, 0.11854765564203262, -0.1486186385154724, 0.2020425945520401, 0.016032667830586433, 0.1181359589099884, -0.09534750878810883, 0.04896572232246399, 0.010238952934741974, -0.010511742904782295, 0.10126926749944687, -0.0035287251230329275, -0.06052067130804062, -0.15023508667945862, -0.11579663306474686, 0.03546925634145737, 0.1182500347495079, -0.061180055141448975, 0.06718824058771133, -0.030428167432546616, 0.0059468927793204784, 0.022467879578471184, -0.059309400618076324, -0.17717739939689636, -0.21762844920158386, 0.03170168027281761, 0.027581622824072838, 0.03280056267976761, -0.0868101716041565, -0.10716906189918518, -0.06264036893844604, 0.21147948503494263, 0.031569499522447586, -0.037066634744405746, -0.13426601886749268, 0.08310942351818085, 0.1424504965543747, -0.06328147649765015, -0.004789974074810743, 0.02922685630619526, 0.16108860075473785, -0.0010541517985984683, -0.038668442517519, 0.028662946075201035, -0.05727385729551315, -0.11692803353071213, -0.03911898657679558, 0.19120901823043823, 0.04815201833844185, 0.06869501620531082, 0.021989459171891212, 0.011030982248485088, 0.020389460027217865, -0.08051755279302597, 0.04471457004547119, 0.025817859917879105, 0.035971373319625854, 0.062052931636571884, -0.01043236255645752, 0.016549184918403625, -0.07722457498311996, -0.045052219182252884, 0.16502730548381805, 0.20557962357997894, -0.06797109544277191, 0.06948012113571167, 0.052348777651786804, -0.051666148006916046, -0.10648918896913528, 0.02037019096314907, 0.13040591776371002, 0.047642942517995834, 0.04490797221660614, -0.18925002217292786, 0.0744132399559021, 0.12474364042282104, -0.007209100294858217, -0.006170181557536125, -0.314446359872818, -0.11504846066236496, 0.09729281812906265, 0.06819109618663788, -0.08441005647182465, -0.12659503519535065, -0.0541142039000988, -0.06473520398139954, -0.10851383954286575, 0.0457758903503418, -0.05335882678627968, 0.10282722860574722, 0.01152726635336876, 0.0864253044128418, 0.031131146475672722, -0.03705357015132904, 0.1735435128211975, 0.02445363625884056, 0.03314436227083206, -0.046490103006362915, 0.06558869034051895, 0.0594884529709816, -0.05307764559984207, 0.059782933443784714, -0.10416970402002335, 0.03776688128709793, -0.16253305971622467, -0.04240615293383598, -0.041248101741075516, 0.05608174204826355, -0.044678863137960434, -0.034675851464271545, -0.04833392798900604, 0.042536742985248566, 0.06588450819253922, -0.025374507531523705, 0.052622247487306595, 0.017037512734532356, 0.07398253679275513, 0.03933991864323616, 0.10737570375204086, -0.0024357859510928392, -0.13324546813964844, -0.037665896117687225, -0.015645377337932587, 0.04338699206709862, -0.05602964013814926, 0.005885125137865543, 0.1216680184006691, 0.04379509389400482, 0.15321335196495056, -0.007314989343285561, -0.09235449135303497, 0.020654959604144096, 0.05501033365726471, -0.034890539944171906, -0.1939566731452942, -0.04030626639723778, 0.046499140560626984, -0.14219798147678375, -0.038668639957904816, 0.08063752949237823, -0.054268136620521545, -0.017633231356739998, -0.011570442467927933, 0.016357462853193283, -0.03919493779540062, 0.18062029778957367, 0.0006616985192522407, 0.07856390625238419, -0.06668613851070404, 0.09541262686252594, 0.1052841916680336, -0.12330707162618637, 0.07708923518657684, 0.04796261712908745, -0.05791110545396805, -0.01454685628414154, 0.03959117829799652, 0.10240496695041656, 0.042283061891794205, -0.06686768680810928, -0.051723629236221313, -0.11574218422174454, 0.03847585245966911, -0.04383397474884987, 0.01597513072192669, -0.03488316386938095, -0.03930988907814026, 0.01489297952502966, -0.1427524834871292, 0.08935067802667618, 0.07180025428533554, 0.041818272322416306, -0.1353566199541092, 0.0737752616405487, 0.022843267768621445, 0.011720617301762104, 0.006940498016774654, -0.00462507177144289, -0.036862812936306, -0.005546413827687502, -0.13962756097316742, -0.037388455122709274, -0.06766866892576218, 0.019478190690279007, -0.012715530581772327, -0.024808675050735474, -0.0520654059946537, 0.02675761468708515, -0.06092674285173416, -0.08909038454294205, -0.0040518613532185555, 0.08291361480951309, -0.11363301426172256, 0.017479307949543, 0.04092748463153839, -0.10309888422489166, 0.0729294866323471, 0.07034233957529068, 0.04446287825703621, 0.032753702253103256, -0.04174355790019035, -0.012001845985651016, 0.026703570038080215, 0.036083199083805084, 0.04692971333861351, -0.1343981921672821, -0.001762665924616158, 0.003222118830308318, 0.013517872430384159, 0.02559003047645092, 0.018824182450771332, -0.10662610083818436, -0.06473860144615173, -0.08963274955749512, -0.035994600504636765, -0.057829536497592926, 0.07724138349294662, 0.11387880891561508, 0.029985615983605385, 0.14005827903747559, -0.08194692432880402, 0.06439457088708878, -0.19315384328365326, -0.018380967900156975, -0.02761600725352764, -0.011493819765746593, -0.016137046739459038, -0.027931569144129753, 0.0742315873503685, -0.041957687586545944, 0.1223512515425682, -0.028564749285578728, 0.0803913101553917, 0.04345951974391937, -0.0493084192276001, 0.001156424987129867, -0.0040508294478058815, 0.20670177042484283, 0.08804678171873093, -0.013421162962913513, 0.06978914141654968, -0.05435739457607269, 0.048345208168029785, 0.062126148492097855, 0.09408849477767944, 0.1773436814546585, 0.009550433605909348, 0.052003163844347, 0.07844215631484985, -0.13329102098941803, -0.14071595668792725, 0.10533100366592407, -0.026425529271364212, 0.10956679284572601, -0.017828775569796562, 0.16202761232852936, 0.13556307554244995, -0.183839350938797, 0.03881370276212692, -0.04277457296848297, -0.11027122288942337, -0.046140264719724655, -0.09395328909158707, -0.06779516488313675, -0.1107340157032013, 0.02944156900048256, -0.08746646344661713, 0.006790222134441137, 0.05479178950190544, 0.025643588975071907, 0.010885936208069324, 0.16724759340286255, -0.04749419167637825, -0.0067937932908535, 0.10180188715457916, -0.011786292307078838, -0.012588459067046642, -0.07578307390213013, -0.04668198525905609, 0.05267135053873062, -0.004364979453384876, 0.10980375111103058, -0.027458081021904945, -0.03405292332172394, 0.041875600814819336, 0.007239534519612789, -0.07765484601259232, 0.0249390359967947, -0.008784189820289612, 0.06216494366526604, 0.07796000689268112, 0.05616670474410057, -0.007031342945992947, -0.06722006946802139, 0.21606537699699402, -0.03997770696878433, -0.05653511360287666, -0.12516909837722778, 0.07627812027931213, 0.025289978832006454, -0.01569635421037674, 0.06216733157634735, -0.10911358147859573, -0.007384934928268194, 0.11113449931144714, 0.148045614361763, 0.005665527656674385, -0.00998670607805252, -0.014500687830150127, -0.014006564393639565, -0.06541953980922699, 0.08205446600914001, 0.0876137912273407, 0.014516494236886501, -0.03653591871261597, 0.04117545485496521, -0.02548993192613125, -0.06327090412378311, -0.06249440461397171, 0.0946478322148323, 0.019816657528281212, -0.006282896269112825, -0.01126084290444851, 0.14965957403182983, 0.010563484393060207, -0.19336692988872528, 0.01018433552235365, -0.13578104972839355, -0.21819089353084564, -0.0015814518555998802, 0.07784711569547653, 0.011177325621247292, 0.05008113011717796, 0.009369303472340107, -0.0033352370373904705, 0.1052517518401146, 0.016472239047288895, -0.050120025873184204, -0.10957319289445877, 0.09820519387722015, -0.03696747124195099, 0.20631349086761475, 0.0015848447801545262, 0.06603658199310303, 0.10265488177537918, 0.04250539839267731, -0.12705084681510925, 0.026641497388482094, 0.09267934411764145, -0.033550117164850235, 0.07921503484249115, 0.183358833193779, -0.05503386631608009, 0.0924520269036293, 0.07099597156047821, -0.09741933643817902, -0.013653981499373913, -0.11079738289117813, 0.041025713086128235, -0.0878610610961914, 0.04261768236756325, -0.04501679167151451, 0.16028983891010284, 0.20341746509075165, -0.05636102706193924, -0.018928248435258865, -0.049991365522146225, 0.017752276733517647, 0.06735944002866745, 0.12397822737693787, -0.018474863842129707, -0.21029537916183472, -0.010981662198901176, -0.023514775559306145, 0.03324295952916145, -0.23203347623348236, -0.10285048931837082, 0.04890844225883484, -0.06738635152578354, -0.01714431308209896, 0.10988672822713852, 0.0773618146777153, 0.013648567721247673, -0.04837188497185707, -0.1438651829957962, -0.034413136541843414, 0.12573637068271637, -0.1673005372285843, -0.018190843984484673 ]
null
null
transformers
# Rick DialoGPT Model
{"tags": ["conversational"]}
text-generation
BW/TEST
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Rick DialoGPT Model
[ "# Rick DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Rick DialoGPT Model" ]
[ 51, 7 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Rick DialoGPT Model" ]
[ -0.027243174612522125, 0.09208611398935318, -0.005486058536916971, 0.01197603065520525, 0.13312271237373352, -0.0006643096567131579, 0.14875547587871552, 0.13561291992664337, -0.012389403767883778, -0.048079900443553925, 0.13848258554935455, 0.20838283002376556, -0.007769247982650995, 0.06212212145328522, -0.07722679525613785, -0.3253750503063202, 0.05440690368413925, 0.05986349284648895, -0.02559526450932026, 0.11941008269786835, 0.10155656188726425, -0.034638021141290665, 0.07502283155918121, 0.008745936676859856, -0.1460564285516739, 0.011253442615270615, 0.020986590534448624, -0.11265120655298233, 0.11301227658987045, 0.0699501633644104, 0.03311868757009506, 0.044131726026535034, -0.04560676962137222, -0.12763948738574982, 0.04502782225608826, 0.00030866602901369333, -0.04332113638520241, 0.05997459217905998, 0.016281595453619957, -0.09000954777002335, 0.11693226546049118, 0.12603440880775452, -0.01263172086328268, 0.041781701147556305, -0.1548357903957367, -0.004369331523776054, -0.01233562733978033, 0.06789606809616089, 0.06087101250886917, 0.10755407065153122, -0.04065045714378357, 0.11729123443365097, -0.06241777911782265, 0.11526333540678024, 0.1129850223660469, -0.291816771030426, -0.016308816149830818, 0.14326390624046326, 0.043570004403591156, 0.04201141744852066, -0.04241296648979187, 0.09895236790180206, 0.01734745316207409, -0.009189855307340622, -0.04667704179883003, -0.07920589298009872, -0.0809992179274559, 0.022899743169546127, -0.08393258601427078, -0.009693359956145287, 0.24909301102161407, -0.033697742968797684, 0.07867740839719772, -0.07909003645181656, -0.08747624605894089, -0.011933685280382633, -0.03604159876704216, -0.03430533409118652, -0.10349667817354202, 0.07883962988853455, -0.03785189241170883, -0.09532928466796875, -0.11454451829195023, -0.029063701629638672, -0.16551746428012848, 0.1769428551197052, 0.028738701716065407, 0.03337583318352699, -0.22648879885673523, 0.09508261829614639, -0.012410550378262997, -0.09879330545663834, 0.018604513257741928, -0.08811058849096298, 0.012304049916565418, 0.017966609448194504, -0.025972042232751846, -0.002111254259943962, 0.08367783576250076, 0.11593183130025864, 0.01627914048731327, 0.018418017774820328, -0.01303142961114645, 0.05024925619363785, 0.039101485162973404, 0.07016518712043762, -0.018131986260414124, -0.026958800852298737, 0.025394905358552933, -0.09519384801387787, -0.01311302836984396, -0.06533002108335495, -0.19878731667995453, -0.008748088963329792, 0.05362382158637047, 0.059645626693964005, 0.040223345160484314, 0.1349429488182068, 0.005914759822189808, -0.04811347648501396, 0.041568055748939514, -0.017372997477650642, -0.016568226739764214, 0.013325352221727371, 0.004558354616165161, 0.14832930266857147, 0.012210249900817871, 0.05107790604233742, -0.11448643356561661, 0.0074756252579391, -0.04443434625864029, -0.019875049591064453, -0.033431850373744965, -0.05190093815326691, -0.010580608621239662, -0.024629589170217514, 0.015543424524366856, -0.1382266879081726, -0.1671048104763031, -0.0113193579018116, -0.006982414051890373, -0.04376089945435524, -0.11932645738124847, -0.1048901304602623, -0.03145192563533783, 0.04379252344369888, -0.060927584767341614, -0.0003760824038181454, -0.04660411551594734, 0.09378229826688766, -0.03543102741241455, 0.07682112604379654, -0.10023638606071472, 0.0828537717461586, -0.07001189142465591, -0.04422231763601303, -0.0734889879822731, 0.13164658844470978, 0.014363138936460018, 0.05487450957298279, -0.031934577971696854, -0.01827416382730007, -0.10224048048257828, 0.07911752909421921, -0.04339373856782913, 0.23623128235340118, -0.09449771791696548, -0.10362883657217026, 0.26979705691337585, -0.053989510983228683, -0.1375254988670349, 0.10795111209154129, -0.015854641795158386, 0.11475867033004761, 0.12686948478221893, 0.18240338563919067, 0.06434911489486694, 0.007867260836064816, 0.07431085407733917, 0.11333738267421722, -0.0774611383676529, -0.018117602914571762, 0.014873803593218327, -0.020292608067393303, -0.07848027348518372, 0.023533256724476814, 0.07671299576759338, 0.05307117849588394, -0.05429181456565857, -0.015286878682672977, 0.00432937266305089, 0.004517627414315939, 0.05698307976126671, -0.02530503273010254, 0.12313884496688843, -0.029461434110999107, -0.07295558601617813, -0.029503753408789635, 0.027530280873179436, -0.05828499048948288, 0.03278997913002968, -0.08230485767126083, 0.03637091815471649, -0.014406797476112843, 0.07024850696325302, -0.16572508215904236, -0.09323301911354065, -0.05250932276248932, 0.1899155229330063, 0.06807822734117508, 0.11413464695215225, 0.05567482113838196, -0.06841246038675308, -0.0038719952572137117, 0.018287649378180504, 0.1991138458251953, -0.01677977479994297, -0.07748494297266006, -0.09769339859485626, 0.10122697055339813, -0.07130109518766403, 0.06141059845685959, -0.050490207970142365, 0.017946461215615273, 0.020556224510073662, 0.1050461083650589, -0.03456922993063927, 0.039413414895534515, 0.011159577406942844, -0.034563858062028885, -0.06218598783016205, -0.004433273337781429, 0.09716981649398804, 0.0021626276429742575, -0.10631977766752243, 0.24286337196826935, -0.19168923795223236, 0.12176351994276047, 0.17641966044902802, -0.19923987984657288, -0.0002552573860157281, -0.11963175982236862, -0.026344671845436096, 0.011637656949460506, 0.037626978009939194, -0.042151857167482376, 0.24314165115356445, -0.00910688005387783, 0.16631373763084412, -0.03389734402298927, -0.04332707077264786, -0.041059546172618866, -0.046011339873075485, 0.010055569931864738, 0.11430004984140396, 0.1047205775976181, -0.17159950733184814, 0.17967921495437622, 0.05867021903395653, 0.05177219957113266, 0.16841758787631989, 0.018001655116677284, 0.021052619442343712, 0.06948674470186234, -0.003431870136409998, -0.03584783151745796, -0.07413756102323532, -0.2106374204158783, -0.023212855681777, 0.0793403834104538, 0.048357341438531876, 0.1068209707736969, -0.1037900522351265, -0.03368109092116356, -0.010547412559390068, -0.021230356767773628, 0.03035620041191578, 0.14086326956748962, 0.013085569255053997, 0.1286563277244568, -0.024180158972740173, -0.06866493821144104, 0.06965550780296326, 0.014881031587719917, -0.08571527898311615, 0.19352088868618011, -0.10702410340309143, -0.34334462881088257, -0.10363983362913132, -0.18596062064170837, -0.056601256132125854, 0.04553624242544174, 0.11461924016475677, -0.14119702577590942, -0.020731983706355095, 0.006813736632466316, 0.06912991404533386, -0.11165751516819, 0.01017086487263441, -0.03630850836634636, -0.017619650810956955, -0.13406261801719666, -0.1034051924943924, -0.05356309190392494, -0.044913630932569504, -0.05510649085044861, 0.12040390819311142, -0.15435875952243805, 0.020806124433875084, 0.23555229604244232, 0.06075655668973923, 0.07018083333969116, -0.03907359018921852, 0.17685799300670624, -0.1052674949169159, 0.011976814828813076, 0.2128676474094391, -0.03831172361969948, 0.06525631994009018, 0.11611197143793106, -0.01394710224121809, -0.0662488266825676, 0.036592915654182434, -0.009823341853916645, -0.07247381657361984, -0.21345274150371552, -0.1158827692270279, -0.1087421104311943, 0.054685093462467194, 0.04713849350810051, 0.050020426511764526, 0.1613347977399826, 0.07427749037742615, -0.04962149262428284, -0.0022197163198143244, 0.06106492131948471, 0.0832381621003151, 0.2504972517490387, -0.06253999471664429, 0.1427627056837082, -0.025090228766202927, -0.16789253056049347, 0.06259234994649887, 0.0661388710141182, 0.09291604906320572, 0.06118352338671684, 0.10224727541208267, 0.005179570056498051, 0.009344357997179031, 0.12825439870357513, 0.07115643471479416, 0.008030776865780354, -0.03595518320798874, -0.039997417479753494, -0.03642706945538521, -0.013250070624053478, 0.032193150371313095, 0.046790316700935364, -0.16567666828632355, -0.021018991246819496, 0.009807335212826729, 0.05824935808777809, 0.02185324765741825, 0.08615364134311676, -0.18498282134532928, -0.016169089823961258, 0.06576614826917648, -0.011832303367555141, -0.11644340306520462, 0.08480028808116913, 0.0007836486911401153, -0.1121063381433487, 0.03723234683275223, -0.027525627985596657, 0.13150714337825775, -0.08457524329423904, 0.0741792693734169, -0.12022519111633301, -0.0374552421271801, -0.010245736688375473, 0.12193918228149414, -0.29501426219940186, 0.19123348593711853, -0.009575535543262959, -0.04439779743552208, -0.1071409061551094, -0.015645509585738182, 0.02963484264910221, 0.10361164063215256, 0.11110331863164902, -0.020523378625512123, -0.02764100395143032, 0.06007368490099907, -0.07205203175544739, 0.0399978905916214, 0.09906689822673798, -0.06730470806360245, -0.013155711814761162, -0.052545808255672455, 0.00039069546619430184, 0.010376452468335629, -0.10966821759939194, 0.022783124819397926, -0.19194799661636353, 0.08703918755054474, 0.08162695169448853, 0.09630028903484344, 0.037212129682302475, -0.029887177050113678, -0.07769683748483658, 0.2589099109172821, 0.009560960344970226, -0.10013746470212936, -0.10953836888074875, 0.008171502500772476, 0.04785030707716942, -0.07699282467365265, -0.016966527327895164, -0.0694924145936966, 0.04450516775250435, -0.06552471220493317, -0.18611730635166168, 0.11722762882709503, -0.09691806137561798, -0.03250948712229729, -0.036249466240406036, 0.21333028376102448, -0.03155504912137985, 0.017869247123599052, 0.04537748545408249, -0.00578570831567049, -0.11741422116756439, -0.10654788464307785, 0.0012778750387951732, -0.004119161982089281, 0.016931969672441483, 0.023226622492074966, -0.03199922665953636, -0.009455137886106968, -0.06797713041305542, -0.014383019879460335, 0.3228513300418854, 0.12615877389907837, -0.042267147451639175, 0.15242800116539001, 0.09877358376979828, -0.06251336634159088, -0.2941497564315796, -0.11165541410446167, -0.07421603053808212, -0.05438753217458725, -0.09733224660158157, -0.18137554824352264, 0.08739634603261948, -0.05383281409740448, -0.013516134582459927, 0.09413999319076538, -0.25194358825683594, -0.10185287892818451, 0.2005643993616104, -0.03753361105918884, 0.4304826855659485, -0.11250142753124237, -0.07815388590097427, -0.04850279167294502, -0.14005880057811737, 0.19035954773426056, 0.004324326757341623, 0.10461755096912384, -0.0006430890643969178, 0.19764995574951172, 0.05591731518507004, -0.0006032987730577588, 0.07056128233671188, 0.01866593211889267, -0.057801030576229095, -0.09095179289579391, -0.0913778692483902, -0.0337459035217762, 0.010270410217344761, 0.0292131919413805, -0.07448325306177139, 0.04388400912284851, -0.13094636797904968, -0.05198022723197937, -0.08626694977283478, 0.038746368139982224, 0.027130719274282455, -0.06653520464897156, -0.0030553280375897884, -0.04914497584104538, 0.0004573945188894868, 0.007742773275822401, 0.21047258377075195, -0.10902713984251022, 0.1467881053686142, 0.028732312843203545, 0.1500566452741623, -0.09794784337282181, -0.04768699035048485, -0.06421241164207458, -0.05478411167860031, 0.07145597785711288, -0.12202182412147522, 0.03240978345274925, 0.1044924184679985, -0.026888413354754448, 0.08732181787490845, 0.1105954647064209, -0.010995322838425636, 0.005803761538118124, 0.08983830362558365, -0.241703063249588, -0.06713853776454926, -0.08410414308309555, 0.05373041704297066, 0.05893997475504875, 0.10275863856077194, 0.20927143096923828, 0.007167487405240536, -0.031165437772870064, 0.021489497274160385, 0.027375908568501472, -0.017840299755334854, 0.05977841466665268, 0.010519524104893208, 0.030491052195429802, -0.14741286635398865, 0.043485816568136215, -0.013757874257862568, -0.09077676385641098, 0.02600322663784027, 0.14754873514175415, -0.10901660472154617, -0.12182232737541199, -0.03921690955758095, 0.13600249588489532, -0.14775370061397552, -0.009947444312274456, -0.0477454848587513, -0.12692049145698547, 0.06857728958129883, 0.1067143976688385, 0.0457911379635334, 0.04121949151158333, -0.09239879250526428, -0.027268609032034874, -0.0535728819668293, 0.00003198942795279436, 0.028995376080274582, -0.0204177163541317, -0.05248761177062988, 0.040780652314424515, -0.03588524088263512, 0.12051229178905487, -0.08552545309066772, -0.10064204037189484, -0.16698434948921204, 0.03528384119272232, -0.07174701243638992, -0.08977310359477997, -0.0871967226266861, -0.03724304214119911, 0.006766482722014189, -0.0405125692486763, -0.02825779654085636, -0.03461418300867081, -0.1126255914568901, 0.03079685941338539, -0.04579872637987137, 0.003088617930188775, -0.07116411626338959, 0.029772473499178886, 0.0525958277285099, -0.029091687873005867, 0.149556964635849, 0.14025014638900757, -0.11192594468593597, 0.09547203034162521, -0.1507159322500229, -0.07066365331411362, 0.09605675935745239, 0.018403515219688416, 0.04981891065835953, 0.05175008252263069, 0.009065150283277035, 0.051755502820014954, 0.06169715151190758, 0.04307684674859047, 0.0153890922665596, -0.07590135186910629, 0.06697173416614532, -0.06090308725833893, -0.10307016223669052, -0.05066140368580818, -0.003966273739933968, 0.015159476548433304, 0.07283487915992737, 0.10097057372331619, -0.056661296635866165, 0.09506311267614365, -0.05649305135011673, 0.04625694453716278, 0.024318000301718712, -0.17797043919563293, 0.03397766128182411, -0.08718447387218475, 0.05030312016606331, 0.010050542652606964, 0.1727033108472824, 0.02054430916905403, -0.019508427008986473, 0.02473587542772293, 0.0719463899731636, 0.04261681064963341, -0.013226886279881, 0.19012948870658875, 0.10657399147748947, -0.03943915665149689, -0.0805516242980957, 0.09759991616010666, 0.04438556358218193, 0.04173632711172104, 0.14543114602565765, -0.05563090741634369, -0.03441290557384491, 0.081944540143013, -0.0026839920319616795, 0.010976077988743782, -0.09896437078714371, -0.13543705642223358, -0.026787811890244484, 0.036508288234472275, -0.03667739778757095, 0.10571453720331192, 0.15851758420467377, -0.005720720160752535, 0.01726081222295761, -0.01855739764869213, -0.05729815363883972, -0.1993623524904251, -0.19528920948505402, -0.083323635160923, -0.13647840917110443, 0.0050200955010950565, -0.13574683666229248, 0.04266147315502167, 0.026296362280845642, 0.09698255360126495, -0.04634363576769829, 0.050944969058036804, 0.03791060671210289, -0.11099781841039658, 0.058360110968351364, -0.043620482087135315, 0.09173028916120529, -0.03267880156636238, 0.014702340587973595, -0.060175783932209015, 0.035412851721048355, 0.016039982438087463, 0.041373249143362045, -0.02921622060239315, 0.019025372341275215, -0.12458328902721405, -0.08709227293729782, -0.06697598844766617, 0.06596853584051132, 0.006195025984197855, 0.16954803466796875, 0.019531596451997757, -0.027915386483073235, 0.028833186253905296, 0.23899038136005402, -0.07318265736103058, -0.09635625779628754, -0.06982157379388809, 0.21012257039546967, -0.009315763600170612, 0.08784335851669312, -0.03747710958123207, 0.009438461624085903, -0.08562079071998596, 0.3506644368171692, 0.29213622212409973, -0.09391074627637863, 0.010968702845275402, -0.0027621579356491566, 0.04181644320487976, 0.12788556516170502, 0.09239348024129868, 0.10824161767959595, 0.29070642590522766, -0.06708572804927826, -0.03647898510098457, -0.006994254421442747, -0.0254643727093935, -0.055716969072818756, 0.0551714263856411, 0.05315792188048363, -0.06511329114437103, -0.01592782698571682, 0.11738577485084534, -0.2489209920167923, 0.0614120177924633, -0.15840938687324524, -0.16190756857395172, -0.07126864790916443, -0.0001230158086400479, 0.0958227664232254, 0.01604771800339222, 0.09578458964824677, -0.011418631300330162, -0.06834693253040314, 0.04414822906255722, 0.020037546753883362, -0.20774760842323303, 0.009963343851268291, 0.06968449801206589, -0.051950447261333466, -0.05526239052414894, -0.017540784552693367, 0.07181108742952347, 0.0862373560667038, 0.031932324171066284, -0.021655123680830002, 0.04088883846998215, -0.011214682832360268, -0.07533704489469528, 0.03916772082448006, 0.027806051075458527, 0.005651058629155159, -0.08518505096435547, 0.07656224071979523, -0.16369622945785522, 0.03412613272666931, -0.0035786160733550787, -0.048953261226415634, -0.014727948233485222, 0.030175231397151947, -0.061420172452926636, 0.08509553223848343, 0.0839199498295784, -0.0171944722533226, -0.016525855287909508, -0.0222842525690794, -0.012990890070796013, -0.020874707028269768, -0.0818524956703186, -0.09698375314474106, -0.15574125945568085, -0.1261346936225891, 0.08575325459241867, -0.00355695397593081, -0.19997835159301758, 0.028783639892935753, -0.12125882506370544, 0.04249454662203789, -0.12142720073461533, 0.09701541811227798, 0.0825105607509613, 0.02303435280919075, -0.0030652873683720827, 0.006164520047605038, 0.03737448528409004, 0.07968182861804962, -0.13731823861598969, -0.08554888516664505 ]
null
null
transformers
[![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/rebel-relation-extraction-by-end-to-end/relation-extraction-on-nyt)](https://paperswithcode.com/sota/relation-extraction-on-nyt?p=rebel-relation-extraction-by-end-to-end) [![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/rebel-relation-extraction-by-end-to-end/relation-extraction-on-conll04)](https://paperswithcode.com/sota/relation-extraction-on-conll04?p=rebel-relation-extraction-by-end-to-end) [![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/rebel-relation-extraction-by-end-to-end/joint-entity-and-relation-extraction-on-3)](https://paperswithcode.com/sota/joint-entity-and-relation-extraction-on-3?p=rebel-relation-extraction-by-end-to-end) [![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/rebel-relation-extraction-by-end-to-end/relation-extraction-on-ade-corpus)](https://paperswithcode.com/sota/relation-extraction-on-ade-corpus?p=rebel-relation-extraction-by-end-to-end) [![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/rebel-relation-extraction-by-end-to-end/relation-extraction-on-re-tacred)](https://paperswithcode.com/sota/relation-extraction-on-re-tacred?p=rebel-relation-extraction-by-end-to-end) ## Multilingual update! Check [mREBEL](https://huggingface.co/Babelscape/mrebel-large), a multilingual version covering more relation types, languages and including entity types. # REBEL <img src="https://i.ibb.co/qsLzNqS/hf-rebel.png" width="30" alt="hf-rebel" border="0" style="display:inline; white-space:nowrap;">: Relation Extraction By End-to-end Language generation This is the model card for the Findings of EMNLP 2021 paper [REBEL: Relation Extraction By End-to-end Language generation](https://github.com/Babelscape/rebel/blob/main/docs/EMNLP_2021_REBEL__Camera_Ready_.pdf). We present a new linearization approach and a reframing of Relation Extraction as a seq2seq task. The paper can be found [here](https://github.com/Babelscape/rebel/blob/main/docs/EMNLP_2021_REBEL__Camera_Ready_.pdf). If you use the code, please reference this work in your paper: @inproceedings{huguet-cabot-navigli-2021-rebel-relation, title = "{REBEL}: Relation Extraction By End-to-end Language generation", author = "Huguet Cabot, Pere-Llu{\'\i}s and Navigli, Roberto", booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021", month = nov, year = "2021", address = "Punta Cana, Dominican Republic", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2021.findings-emnlp.204", pages = "2370--2381", abstract = "Extracting relation triplets from raw text is a crucial task in Information Extraction, enabling multiple applications such as populating or validating knowledge bases, factchecking, and other downstream tasks. However, it usually involves multiple-step pipelines that propagate errors or are limited to a small number of relation types. To overcome these issues, we propose the use of autoregressive seq2seq models. Such models have previously been shown to perform well not only in language generation, but also in NLU tasks such as Entity Linking, thanks to their framing as seq2seq tasks. In this paper, we show how Relation Extraction can be simplified by expressing triplets as a sequence of text and we present REBEL, a seq2seq model based on BART that performs end-to-end relation extraction for more than 200 different relation types. We show our model{'}s flexibility by fine-tuning it on an array of Relation Extraction and Relation Classification benchmarks, with it attaining state-of-the-art performance in most of them.", } The original repository for the paper can be found [here](https://github.com/Babelscape/rebel) Be aware that the inference widget at the right does not output special tokens, which are necessary to distinguish the subject, object and relation types. For a demo of REBEL and its pre-training dataset check the [Spaces demo](https://huggingface.co/spaces/Babelscape/rebel-demo). ## Pipeline usage ```python from transformers import pipeline triplet_extractor = pipeline('text2text-generation', model='Babelscape/rebel-large', tokenizer='Babelscape/rebel-large') # We need to use the tokenizer manually since we need special tokens. extracted_text = triplet_extractor.tokenizer.batch_decode([triplet_extractor("Punta Cana is a resort town in the municipality of Higuey, in La Altagracia Province, the eastern most province of the Dominican Republic", return_tensors=True, return_text=False)[0]["generated_token_ids"]]) print(extracted_text[0]) # Function to parse the generated text and extract the triplets def extract_triplets(text): triplets = [] relation, subject, relation, object_ = '', '', '', '' text = text.strip() current = 'x' for token in text.replace("<s>", "").replace("<pad>", "").replace("</s>", "").split(): if token == "<triplet>": current = 't' if relation != '': triplets.append({'head': subject.strip(), 'type': relation.strip(),'tail': object_.strip()}) relation = '' subject = '' elif token == "<subj>": current = 's' if relation != '': triplets.append({'head': subject.strip(), 'type': relation.strip(),'tail': object_.strip()}) object_ = '' elif token == "<obj>": current = 'o' relation = '' else: if current == 't': subject += ' ' + token elif current == 's': object_ += ' ' + token elif current == 'o': relation += ' ' + token if subject != '' and relation != '' and object_ != '': triplets.append({'head': subject.strip(), 'type': relation.strip(),'tail': object_.strip()}) return triplets extracted_triplets = extract_triplets(extracted_text[0]) print(extracted_triplets) ``` ## Model and Tokenizer using transformers ```python from transformers import AutoModelForSeq2SeqLM, AutoTokenizer def extract_triplets(text): triplets = [] relation, subject, relation, object_ = '', '', '', '' text = text.strip() current = 'x' for token in text.replace("<s>", "").replace("<pad>", "").replace("</s>", "").split(): if token == "<triplet>": current = 't' if relation != '': triplets.append({'head': subject.strip(), 'type': relation.strip(),'tail': object_.strip()}) relation = '' subject = '' elif token == "<subj>": current = 's' if relation != '': triplets.append({'head': subject.strip(), 'type': relation.strip(),'tail': object_.strip()}) object_ = '' elif token == "<obj>": current = 'o' relation = '' else: if current == 't': subject += ' ' + token elif current == 's': object_ += ' ' + token elif current == 'o': relation += ' ' + token if subject != '' and relation != '' and object_ != '': triplets.append({'head': subject.strip(), 'type': relation.strip(),'tail': object_.strip()}) return triplets # Load model and tokenizer tokenizer = AutoTokenizer.from_pretrained("Babelscape/rebel-large") model = AutoModelForSeq2SeqLM.from_pretrained("Babelscape/rebel-large") gen_kwargs = { "max_length": 256, "length_penalty": 0, "num_beams": 3, "num_return_sequences": 3, } # Text to extract triplets from text = 'Punta Cana is a resort town in the municipality of Higüey, in La Altagracia Province, the easternmost province of the Dominican Republic.' # Tokenizer text model_inputs = tokenizer(text, max_length=256, padding=True, truncation=True, return_tensors = 'pt') # Generate generated_tokens = model.generate( model_inputs["input_ids"].to(model.device), attention_mask=model_inputs["attention_mask"].to(model.device), **gen_kwargs, ) # Extract text decoded_preds = tokenizer.batch_decode(generated_tokens, skip_special_tokens=False) # Extract triplets for idx, sentence in enumerate(decoded_preds): print(f'Prediction triplets sentence {idx}') print(extract_triplets(sentence)) ```
{"language": ["en"], "license": "cc-by-nc-sa-4.0", "tags": ["seq2seq", "relation-extraction"], "datasets": ["Babelscape/rebel-dataset"], "widget": [{"text": "Punta Cana is a resort town in the municipality of Higuey, in La Altagracia Province, the eastern most province of the Dominican Republic"}], "model-index": [{"name": "REBEL", "results": [{"task": {"type": "Relation-Extraction", "name": "Relation Extraction"}, "dataset": {"name": "CoNLL04", "type": "CoNLL04"}, "metrics": [{"type": "re+ macro f1", "value": 76.65, "name": "RE+ Macro F1"}]}, {"task": {"type": "Relation-Extraction", "name": "Relation Extraction"}, "dataset": {"name": "NYT", "type": "NYT"}, "metrics": [{"type": "f1", "value": 93.4, "name": "F1"}]}]}]}
text2text-generation
Babelscape/rebel-large
[ "transformers", "pytorch", "safetensors", "bart", "text2text-generation", "seq2seq", "relation-extraction", "en", "dataset:Babelscape/rebel-dataset", "license:cc-by-nc-sa-4.0", "model-index", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #safetensors #bart #text2text-generation #seq2seq #relation-extraction #en #dataset-Babelscape/rebel-dataset #license-cc-by-nc-sa-4.0 #model-index #autotrain_compatible #endpoints_compatible #has_space #region-us
![PWC](URL ![PWC](URL ![PWC](URL ![PWC](URL ![PWC](URL ## Multilingual update! Check mREBEL, a multilingual version covering more relation types, languages and including entity types. # REBEL <img src="https://i.URL width="30" alt="hf-rebel" border="0" style="display:inline; white-space:nowrap;">: Relation Extraction By End-to-end Language generation This is the model card for the Findings of EMNLP 2021 paper REBEL: Relation Extraction By End-to-end Language generation. We present a new linearization approach and a reframing of Relation Extraction as a seq2seq task. The paper can be found here. If you use the code, please reference this work in your paper: @inproceedings{huguet-cabot-navigli-2021-rebel-relation, title = "{REBEL}: Relation Extraction By End-to-end Language generation", author = "Huguet Cabot, Pere-Llu{\'\i}s and Navigli, Roberto", booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021", month = nov, year = "2021", address = "Punta Cana, Dominican Republic", publisher = "Association for Computational Linguistics", url = "URL pages = "2370--2381", abstract = "Extracting relation triplets from raw text is a crucial task in Information Extraction, enabling multiple applications such as populating or validating knowledge bases, factchecking, and other downstream tasks. However, it usually involves multiple-step pipelines that propagate errors or are limited to a small number of relation types. To overcome these issues, we propose the use of autoregressive seq2seq models. Such models have previously been shown to perform well not only in language generation, but also in NLU tasks such as Entity Linking, thanks to their framing as seq2seq tasks. In this paper, we show how Relation Extraction can be simplified by expressing triplets as a sequence of text and we present REBEL, a seq2seq model based on BART that performs end-to-end relation extraction for more than 200 different relation types. We show our model{'}s flexibility by fine-tuning it on an array of Relation Extraction and Relation Classification benchmarks, with it attaining state-of-the-art performance in most of them.", } The original repository for the paper can be found here Be aware that the inference widget at the right does not output special tokens, which are necessary to distinguish the subject, object and relation types. For a demo of REBEL and its pre-training dataset check the Spaces demo. ## Pipeline usage ## Model and Tokenizer using transformers
[ "## Multilingual update! Check mREBEL, a multilingual version covering more relation types, languages and including entity types.", "# REBEL <img src=\"https://i.URL width=\"30\" alt=\"hf-rebel\" border=\"0\" style=\"display:inline; white-space:nowrap;\">: Relation Extraction By End-to-end Language generation\nThis is the model card for the Findings of EMNLP 2021 paper REBEL: Relation Extraction By End-to-end Language generation. We present a new linearization approach and a reframing of Relation Extraction as a seq2seq task. The paper can be found here. If you use the code, please reference this work in your paper:\n\n @inproceedings{huguet-cabot-navigli-2021-rebel-relation,\n title = \"{REBEL}: Relation Extraction By End-to-end Language generation\",\n author = \"Huguet Cabot, Pere-Llu{\\'\\i}s and\n Navigli, Roberto\",\n booktitle = \"Findings of the Association for Computational Linguistics: EMNLP 2021\",\n month = nov,\n year = \"2021\",\n address = \"Punta Cana, Dominican Republic\",\n publisher = \"Association for Computational Linguistics\",\n url = \"URL\n pages = \"2370--2381\",\n abstract = \"Extracting relation triplets from raw text is a crucial task in Information Extraction, enabling multiple applications such as populating or validating knowledge bases, factchecking, and other downstream tasks. However, it usually involves multiple-step pipelines that propagate errors or are limited to a small number of relation types. To overcome these issues, we propose the use of autoregressive seq2seq models. Such models have previously been shown to perform well not only in language generation, but also in NLU tasks such as Entity Linking, thanks to their framing as seq2seq tasks. In this paper, we show how Relation Extraction can be simplified by expressing triplets as a sequence of text and we present REBEL, a seq2seq model based on BART that performs end-to-end relation extraction for more than 200 different relation types. We show our model{'}s flexibility by fine-tuning it on an array of Relation Extraction and Relation Classification benchmarks, with it attaining state-of-the-art performance in most of them.\",\n }\n\nThe original repository for the paper can be found here\n\nBe aware that the inference widget at the right does not output special tokens, which are necessary to distinguish the subject, object and relation types. For a demo of REBEL and its pre-training dataset check the Spaces demo.", "## Pipeline usage", "## Model and Tokenizer using transformers" ]
[ "TAGS\n#transformers #pytorch #safetensors #bart #text2text-generation #seq2seq #relation-extraction #en #dataset-Babelscape/rebel-dataset #license-cc-by-nc-sa-4.0 #model-index #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "## Multilingual update! Check mREBEL, a multilingual version covering more relation types, languages and including entity types.", "# REBEL <img src=\"https://i.URL width=\"30\" alt=\"hf-rebel\" border=\"0\" style=\"display:inline; white-space:nowrap;\">: Relation Extraction By End-to-end Language generation\nThis is the model card for the Findings of EMNLP 2021 paper REBEL: Relation Extraction By End-to-end Language generation. We present a new linearization approach and a reframing of Relation Extraction as a seq2seq task. The paper can be found here. If you use the code, please reference this work in your paper:\n\n @inproceedings{huguet-cabot-navigli-2021-rebel-relation,\n title = \"{REBEL}: Relation Extraction By End-to-end Language generation\",\n author = \"Huguet Cabot, Pere-Llu{\\'\\i}s and\n Navigli, Roberto\",\n booktitle = \"Findings of the Association for Computational Linguistics: EMNLP 2021\",\n month = nov,\n year = \"2021\",\n address = \"Punta Cana, Dominican Republic\",\n publisher = \"Association for Computational Linguistics\",\n url = \"URL\n pages = \"2370--2381\",\n abstract = \"Extracting relation triplets from raw text is a crucial task in Information Extraction, enabling multiple applications such as populating or validating knowledge bases, factchecking, and other downstream tasks. However, it usually involves multiple-step pipelines that propagate errors or are limited to a small number of relation types. To overcome these issues, we propose the use of autoregressive seq2seq models. Such models have previously been shown to perform well not only in language generation, but also in NLU tasks such as Entity Linking, thanks to their framing as seq2seq tasks. In this paper, we show how Relation Extraction can be simplified by expressing triplets as a sequence of text and we present REBEL, a seq2seq model based on BART that performs end-to-end relation extraction for more than 200 different relation types. We show our model{'}s flexibility by fine-tuning it on an array of Relation Extraction and Relation Classification benchmarks, with it attaining state-of-the-art performance in most of them.\",\n }\n\nThe original repository for the paper can be found here\n\nBe aware that the inference widget at the right does not output special tokens, which are necessary to distinguish the subject, object and relation types. For a demo of REBEL and its pre-training dataset check the Spaces demo.", "## Pipeline usage", "## Model and Tokenizer using transformers" ]
[ 91, 30, 592, 4, 9 ]
[ "passage: TAGS\n#transformers #pytorch #safetensors #bart #text2text-generation #seq2seq #relation-extraction #en #dataset-Babelscape/rebel-dataset #license-cc-by-nc-sa-4.0 #model-index #autotrain_compatible #endpoints_compatible #has_space #region-us \n## Multilingual update! Check mREBEL, a multilingual version covering more relation types, languages and including entity types." ]
[ -0.09480569511651993, 0.13649532198905945, -0.004141831770539284, 0.08438782393932343, 0.09423216432332993, -0.003795123193413019, 0.13760291039943695, 0.07323440909385681, 0.05451400578022003, -0.04681999236345291, 0.09005756676197052, 0.10873160511255264, -0.010351322591304779, 0.13166414201259613, -0.03296046331524849, -0.25076353549957275, 0.053018759936094284, -0.020454727113246918, 0.04677869379520416, 0.09426401555538177, 0.139424666762352, -0.036327064037323, 0.09298722445964813, -0.0005333814769983292, -0.09588655829429626, 0.033459633588790894, 0.0011535210069268942, -0.10271558165550232, 0.12437207996845245, 0.05633706599473953, 0.08171868324279785, 0.08760560303926468, 0.06776817888021469, -0.16188892722129822, 0.03139672800898552, -0.03437325358390808, -0.053656112402677536, 0.016127584502100945, 0.04220713675022125, -0.09943782538175583, 0.059426046907901764, -0.05586237460374832, -0.005121674854308367, 0.05526365712285042, -0.060558442026376724, -0.07050339877605438, -0.08420903980731964, 0.023529106751084328, 0.0463499091565609, 0.055757343769073486, 0.012375171296298504, 0.15050996840000153, -0.05131552740931511, 0.07572343200445175, 0.10877689719200134, -0.31714996695518494, 0.0064714159816503525, 0.1242712140083313, 0.06121710687875748, 0.0129474438726902, -0.04750637710094452, 0.08175778388977051, 0.05138557031750679, 0.010290954262018204, 0.007868435233831406, -0.11475488543510437, -0.10636863857507706, 0.008817806839942932, -0.09190984815359116, 0.015036358498036861, 0.227777898311615, -0.03236335143446922, -0.03267044201493263, -0.06072814390063286, -0.035918109118938446, 0.11812211573123932, -0.040816530585289, 0.006210798863321543, 0.00878247618675232, 0.014696920290589333, 0.02222173660993576, -0.06210188940167427, -0.1069805920124054, -0.006199283991008997, -0.20640772581100464, 0.13755550980567932, 0.028010500594973564, 0.06278247386217117, -0.06389918178319931, -0.007569675333797932, -0.05257616937160492, -0.10383874177932739, -0.05335242301225662, -0.08433166146278381, -0.004530756268650293, 0.005132706835865974, -0.07780604809522629, -0.0035061959642916918, 0.13920338451862335, 0.11060934513807297, -0.07354125380516052, 0.03376217558979988, -0.05910637602210045, 0.05929518863558769, 0.04909691959619522, 0.07240425050258636, -0.01919066347181797, -0.09526028484106064, 0.10511139035224915, -0.05632156506180763, 0.05805454030632973, -0.027155881747603416, -0.17073234915733337, -0.06923873722553253, -0.04080981761217117, 0.11369141191244125, -0.011129127815365791, 0.09921079874038696, -0.0025111909490078688, 0.017123626545071602, 0.08812490850687027, -0.11522846668958664, 0.038257598876953125, 0.008071091957390308, 0.030316969379782677, 0.012625837698578835, 0.00968789029866457, 0.029944350942969322, -0.0571366511285305, -0.02367229387164116, -0.024404466152191162, -0.0048788851127028465, -0.04933981969952583, -0.08191261440515518, 0.07719718664884567, -0.0015657780459150672, 0.0020169918425381184, -0.1804843693971634, -0.2102796882390976, 0.023258361965417862, -0.012139620259404182, -0.005991288926452398, -0.03866422921419144, -0.04909105226397514, -0.008961914107203484, 0.023507773876190186, -0.05924665927886963, -0.06706646829843521, -0.06496357917785645, 0.00216105068102479, -0.013795559294521809, 0.034094683825969696, -0.1827106773853302, 0.0043769702315330505, -0.11468417942523956, -0.01680910214781761, -0.04984861612319946, 0.0870227962732315, -0.08923105895519257, 0.013964188285171986, -0.040652986615896225, -0.033648062497377396, -0.01580905169248581, 0.06698821485042572, -0.042275965213775635, 0.16981405019760132, -0.24329611659049988, -0.05075570568442345, 0.230869323015213, -0.1418435126543045, -0.15775135159492493, 0.1195000633597374, -0.0223537627607584, -0.06365291029214859, 0.08791279792785645, 0.13677746057510376, 0.09911192208528519, -0.02275260165333748, -0.017573462799191475, 0.09749718010425568, -0.007950607687234879, -0.08494502305984497, 0.07368957251310349, 0.010310729034245014, -0.05177099257707596, 0.058722175657749176, -0.012622665613889694, 0.05033772066235542, -0.006623400375247002, -0.06241879239678383, -0.0325145460665226, -0.01732318289577961, 0.06104052811861038, -0.008394373580813408, 0.04579859972000122, -0.08853405714035034, -0.006192485336214304, 0.020718220621347427, 0.07023780792951584, -0.010334191843867302, 0.038588158786296844, -0.07996708899736404, 0.04064870625734329, 0.03777269646525383, 0.0352477990090847, -0.13849332928657532, -0.09582769870758057, -0.01874723844230175, 0.09267596900463104, 0.08703741431236267, 0.04571044072508812, 0.02870839089155197, 0.01640450209379196, -0.04779116436839104, 0.03590782731771469, 0.09553300589323044, 0.0098077692091465, -0.006889474578201771, -0.1447390913963318, 0.04309282451868057, -0.09584908932447433, 0.08836954087018967, -0.046083077788352966, 0.020545653998851776, 0.0225205197930336, 0.1301904171705246, -0.032522447407245636, 0.07741041481494904, -0.013951714150607586, 0.06546807289123535, -0.07670623064041138, 0.015597127377986908, 0.06675263494253159, -0.008798778988420963, -0.06720009446144104, 0.24367456138134003, -0.1030358150601387, 0.18408317863941193, 0.20116335153579712, -0.03860560804605484, 0.039747994393110275, -0.11121923476457596, 0.016407763585448265, 0.043627962470054626, 0.013267395086586475, 0.001465281005948782, 0.0825437381863594, 0.04261776804924011, 0.13282178342342377, -0.08347655087709427, 0.0013173951301723719, -0.020219260826706886, -0.07984350621700287, -0.08749710768461227, 0.06893935799598694, 0.07790124416351318, -0.19727979600429535, 0.157823383808136, 0.2440275400876999, -0.00967077724635601, 0.15503722429275513, -0.026566406711935997, 0.03394463658332825, -0.0027716869954019785, -0.05853456258773804, -0.030504344031214714, 0.04235946759581566, -0.05502558872103691, -0.05706493556499481, 0.05108869448304176, 0.01246942300349474, 0.04239766299724579, -0.0989023968577385, -0.03276345133781433, -0.007371042855083942, -0.039260998368263245, -0.03499297797679901, 0.05475452169775963, 0.03737829625606537, 0.08177607506513596, -0.09035392850637436, -0.09974262863397598, 0.028462953865528107, 0.0009449059725739062, -0.09461169689893723, 0.19924655556678772, -0.19179567694664001, -0.2544053792953491, -0.13388559222221375, -0.15058039128780365, -0.13786102831363678, 0.042147595435380936, 0.07648596912622452, -0.025660933926701546, -0.03959457948803902, -0.059457890689373016, 0.020833399146795273, -0.05427027493715286, -0.005132114049047232, 0.011206346563994884, 0.03370228037238121, -0.04002619907259941, -0.11990474164485931, -0.02901424467563629, 0.016648169606924057, -0.08405212312936783, 0.07903677970170975, -0.08394816517829895, 0.10143974423408508, 0.11381339281797409, -0.03270229697227478, 0.013978971168398857, -0.07619471848011017, 0.13562613725662231, 0.001355437794700265, -0.0616377629339695, 0.1783505082130432, 0.004125040024518967, 0.029663003981113434, 0.1454348862171173, 0.045531272888183594, -0.028750533238053322, -0.004368741065263748, -0.01191767118871212, -0.0554901584982872, -0.15716096758842468, -0.15073198080062866, -0.0668044462800026, 0.04795315861701965, -0.02861269935965538, 0.03589813411235809, -0.036208733916282654, 0.050760235637426376, -0.036490652710199356, -0.04406186193227768, 0.057482097297906876, 0.07528367638587952, 0.25336629152297974, -0.05131146311759949, 0.1335831880569458, -0.06338571012020111, -0.07412942498922348, 0.09092025458812714, 0.12000570446252823, 0.07885061949491501, 0.06518439203500748, 0.018757693469524384, 0.07708493620157242, 0.09241826832294464, 0.07267486304044724, 0.0569530725479126, 0.02106650546193123, 0.0026907797437161207, -0.031097738072276115, -0.050112541764974594, -0.04081365838646889, 0.03105446882545948, 0.013238575309515, -0.08973192423582077, -0.08504530787467957, -0.039132364094257355, 0.10350801050662994, 0.020971614867448807, 0.028684554621577263, -0.2037963569164276, 0.019611917436122894, 0.09353359043598175, 0.009142507798969746, -0.04305572807788849, 0.07834582030773163, 0.010999076999723911, -0.08943843096494675, 0.18181177973747253, -0.0016543292440474033, 0.1106024980545044, 0.005741569679230452, 0.030976084992289543, -0.1094701737165451, -0.0039007205050438643, 0.03121868707239628, 0.12283503264188766, -0.27283379435539246, 0.2309679388999939, 0.03444148227572441, -0.029778171330690384, -0.05979021638631821, 0.0037427230272442102, 0.04618173837661743, 0.2395528256893158, 0.10546962916851044, -0.009921658784151077, -0.15179875493049622, -0.09586810320615768, -0.02497388795018196, 0.06337942183017731, 0.056099724024534225, 0.020134001970291138, 0.03846878930926323, -0.02707579731941223, -0.02694263868033886, -0.009990091435611248, 0.1225312128663063, -0.12908287346363068, -0.19533561170101166, 0.027956988662481308, 0.1209336519241333, -0.07355137914419174, -0.044307220727205276, -0.07472587376832962, -0.11854194849729538, 0.13484381139278412, -0.038386985659599304, -0.054637327790260315, -0.10845271497964859, -0.06545600295066833, 0.06614946573972702, -0.09822887182235718, 0.015437323600053787, -0.042788341641426086, 0.10662201046943665, -0.09344929456710815, -0.15187737345695496, 0.0321102999150753, -0.10062461346387863, -0.01831863820552826, -0.04554217681288719, 0.12254414707422256, -0.06582097709178925, -0.007298586890101433, 0.08682917058467865, 0.011917534284293652, -0.024624096229672432, -0.1320386528968811, -0.06557769328355789, 0.05009225010871887, 0.0971280187368393, 0.05057093873620033, -0.1356678158044815, -0.179144024848938, 0.021159978583455086, -0.016477983444929123, 0.22706596553325653, 0.17823155224323273, -0.07375995069742203, 0.12494470179080963, 0.18215899169445038, -0.08335184305906296, -0.3100673258304596, -0.13904684782028198, -0.048255253583192825, 0.024644481018185616, 0.02237333357334137, -0.0972161814570427, 0.0508040152490139, 0.03173438459634781, -0.05123043432831764, -0.09154626727104187, -0.16573496162891388, -0.10072392225265503, 0.1651051640510559, -0.0067319925874471664, 0.16259264945983887, -0.12313012033700943, -0.06405779719352722, -0.1173117533326149, -0.1415896713733673, 0.188076451420784, -0.13822969794273376, 0.05595788732171059, 0.0023671607486903667, 0.0885126143693924, 0.00691090477630496, -0.006847999058663845, 0.12730173766613007, -0.009613746777176857, 0.01944483444094658, -0.06482420861721039, -0.0033974649850279093, 0.0913645476102829, -0.011484586633741856, 0.06263697892427444, -0.17410747706890106, 0.03422662615776062, -0.0857338160276413, -0.05243823304772377, -0.0409383662045002, 0.11412093043327332, -0.01898866891860962, -0.030662449076771736, -0.0024658837355673313, 0.012292467057704926, -0.006170485634356737, -0.04486057162284851, 0.17628377676010132, -0.04808073490858078, 0.046934280544519424, 0.18247608840465546, 0.1427699476480484, -0.14453063905239105, 0.08800464123487473, -0.04547012597322464, -0.08188227564096451, 0.02971521206200123, -0.06042833253741264, 0.04243973270058632, 0.07614807039499283, -0.03996321186423302, 0.06741257756948471, 0.06427283585071564, -0.00018186903616879135, 0.00446705985814333, 0.1371629685163498, -0.09179427474737167, -0.06087110936641693, -0.022986577823758125, 0.03744873031973839, 0.008228025399148464, 0.08955360949039459, 0.18137897551059723, -0.02623078227043152, -0.0439961738884449, -0.01841544732451439, 0.0074080172926187515, -0.030187727883458138, 0.0646858811378479, 0.05474770441651344, 0.01274045743048191, -0.11498914659023285, 0.09046320617198944, 0.03160310909152031, -0.0586446076631546, -0.0009826075984165072, 0.11805623769760132, -0.09809699654579163, -0.13886338472366333, -0.009995301254093647, 0.13559141755104065, -0.20637866854667664, -0.07918420433998108, -0.03995683416724205, -0.12175822257995605, 0.012244096025824547, 0.1887054741382599, 0.08203938603401184, 0.027516480535268784, -0.017562562599778175, -0.07095341384410858, 0.01268300786614418, 0.07010052353143692, 0.0036192545667290688, 0.030339127406477928, -0.10105881094932556, 0.06752914190292358, -0.05632135272026062, 0.11766841262578964, -0.06039568781852722, 0.00982443243265152, -0.10543237626552582, 0.0019832521211355925, -0.19228075444698334, 0.04250209033489227, -0.1312345564365387, 0.01699751242995262, 0.0029835524037480354, -0.05945976823568344, -0.0392957367002964, 0.00738499453291297, -0.07643628865480423, 0.005343222990632057, -0.04381903260946274, 0.11934959143400192, -0.13614781200885773, -0.03792286664247513, 0.050133511424064636, 0.0118588050827384, 0.09328990429639816, 0.06159385293722153, -0.07587552070617676, 0.11272790282964706, -0.2055310308933258, -0.07580860704183578, 0.08384465426206589, 0.08613769710063934, 0.05736056715250015, 0.00905879307538271, 0.025761378929018974, 0.12499826401472092, 0.027598492801189423, 0.05385298281908035, 0.1113990843296051, -0.05228175222873688, -0.05602913349866867, -0.03239873796701431, -0.10361765325069427, 0.0017165003810077906, -0.014315951615571976, 0.11710092425346375, 0.06262974441051483, 0.14602316915988922, -0.06761648505926132, 0.029438793659210205, -0.08653569221496582, 0.024359140545129776, -0.0391443595290184, -0.13637900352478027, -0.13276346027851105, -0.06847364455461502, 0.02816902846097946, 0.006713725160807371, 0.28481659293174744, 0.06525677442550659, 0.05746316537261009, 0.03405435383319855, 0.05339713767170906, 0.06892111152410507, 0.03858247399330139, 0.20226967334747314, 0.0345240943133831, 0.01688382215797901, -0.02125384844839573, 0.022865960374474525, 0.009314451366662979, -0.05633806064724922, 0.04133755341172218, 0.10485129803419113, 0.07242093235254288, 0.09403672814369202, 0.11210189014673233, 0.0075417738407850266, 0.018565300852060318, -0.16246215999126434, 0.011944375932216644, 0.019582148641347885, -0.06748520582914352, 0.08402886986732483, 0.12717002630233765, -0.10163601487874985, 0.04628279060125351, -0.032619770616292953, -0.05963641405105591, -0.17491655051708221, -0.1242930144071579, -0.12439259141683578, -0.10081176459789276, -0.0016920518828555942, -0.14443883299827576, 0.013866450637578964, 0.010689888149499893, 0.043898385018110275, -0.014652984216809273, -0.019479943439364433, -0.09412692487239838, -0.08211097121238708, 0.010307058691978455, 0.009054673835635185, 0.07081928849220276, 0.02935955487191677, -0.02540159411728382, -0.023320775479078293, -0.00869261845946312, -0.020453201606869698, 0.01382377464324236, 0.025193873792886734, 0.04417787864804268, -0.15029209852218628, -0.03272473067045212, -0.02952622063457966, 0.02339976280927658, 0.09657584130764008, 0.14976634085178375, 0.053837746381759644, -0.05742030218243599, 0.06389319151639938, 0.22922053933143616, -0.025089558213949203, -0.15467867255210876, -0.07906953245401382, 0.19913969933986664, 0.02937135472893715, 0.0956769585609436, -0.02538059465587139, -0.03205370157957077, -0.02862953580915928, 0.1823839694261551, 0.30226263403892517, -0.12441915273666382, 0.0331064872443676, -0.055746208876371384, 0.02562759444117546, 0.06865016371011734, 0.09226606041193008, 0.0796680748462677, 0.21861162781715393, -0.017751410603523254, -0.03714204579591751, -0.01697671227157116, 0.04482940584421158, -0.15500079095363617, 0.09614825248718262, -0.07459503412246704, -0.09433021396398544, 0.00018243002705276012, 0.09766171127557755, -0.049983374774456024, 0.056637246161699295, -0.07310906797647476, -0.1380133330821991, -0.07308045774698257, 0.00874129869043827, 0.10029838234186172, 0.02607828937470913, -0.003828680142760277, -0.010752675123512745, 0.005986551754176617, 0.06459706276655197, -0.013321341946721077, -0.15867255628108978, -0.021720558404922485, 0.05735335126519203, -0.0025589182041585445, 0.13646100461483002, 0.011732835322618484, 0.10589885711669922, 0.08736324310302734, 0.05065084248781204, -0.07384911179542542, 0.07965999096632004, 0.019088037312030792, 0.021666130051016808, 0.07069583982229233, -0.01240905188024044, -0.030707726255059242, 0.025460273027420044, 0.059304505586624146, -0.06979348510503769, 0.08675158768892288, 0.06904830783605576, -0.07881885766983032, -0.046786803752183914, 0.11992046982049942, -0.13635306060314178, 0.07027982175350189, 0.07220721244812012, 0.0014200408477336168, -0.023021064698696136, -0.030743027105927467, 0.043635301291942596, 0.0021892243530601263, -0.04536871239542961, -0.05197354778647423, -0.07664299011230469, -0.05496696010231972, 0.037868209183216095, 0.027534298598766327, -0.1749478578567505, -0.040622975677251816, -0.07333090156316757, 0.04452075809240341, -0.08938857913017273, 0.11041634529829025, 0.10854527354240417, -0.04518234357237816, -0.02396593615412712, -0.19202278554439545, 0.06546513736248016, 0.046258557587862015, -0.10935220122337341, -0.09968262165784836 ]
null
null
transformers
# WikiNEuRal: Combined Neural and Knowledge-based Silver Data Creation for Multilingual NER This is the model card for the EMNLP 2021 paper [WikiNEuRal: Combined Neural and Knowledge-based Silver Data Creation for Multilingual NER](https://aclanthology.org/2021.findings-emnlp.215/). We fine-tuned a multilingual language model (mBERT) for 3 epochs on our [WikiNEuRal dataset](https://huggingface.co/datasets/Babelscape/wikineural) for Named Entity Recognition (NER). The resulting multilingual NER model supports the 9 languages covered by WikiNEuRal (de, en, es, fr, it, nl, pl, pt, ru), and it was trained on all 9 languages jointly. **If you use the model, please reference this work in your paper**: ```bibtex @inproceedings{tedeschi-etal-2021-wikineural-combined, title = "{W}iki{NE}u{R}al: {C}ombined Neural and Knowledge-based Silver Data Creation for Multilingual {NER}", author = "Tedeschi, Simone and Maiorca, Valentino and Campolungo, Niccol{\`o} and Cecconi, Francesco and Navigli, Roberto", booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021", month = nov, year = "2021", address = "Punta Cana, Dominican Republic", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2021.findings-emnlp.215", pages = "2521--2533", abstract = "Multilingual Named Entity Recognition (NER) is a key intermediate task which is needed in many areas of NLP. In this paper, we address the well-known issue of data scarcity in NER, especially relevant when moving to a multilingual scenario, and go beyond current approaches to the creation of multilingual silver data for the task. We exploit the texts of Wikipedia and introduce a new methodology based on the effective combination of knowledge-based approaches and neural models, together with a novel domain adaptation technique, to produce high-quality training corpora for NER. We evaluate our datasets extensively on standard benchmarks for NER, yielding substantial improvements up to 6 span-based F1-score points over previous state-of-the-art systems for data creation.", } ``` The original repository for the paper can be found at [https://github.com/Babelscape/wikineural](https://github.com/Babelscape/wikineural). ## How to use You can use this model with Transformers *pipeline* for NER. ```python from transformers import AutoTokenizer, AutoModelForTokenClassification from transformers import pipeline tokenizer = AutoTokenizer.from_pretrained("Babelscape/wikineural-multilingual-ner") model = AutoModelForTokenClassification.from_pretrained("Babelscape/wikineural-multilingual-ner") nlp = pipeline("ner", model=model, tokenizer=tokenizer, grouped_entities=True) example = "My name is Wolfgang and I live in Berlin" ner_results = nlp(example) print(ner_results) ``` ## Limitations and bias This model is trained on WikiNEuRal, a state-of-the-art dataset for Multilingual NER automatically derived from Wikipedia. Therefore, it might not generalize well to all textual genres (e.g. news). On the other hand, models trained only on news articles (e.g. only on CoNLL03) have been proven to obtain much lower scores on encyclopedic articles. To obtain more robust systems, we encourage you to train a system on the combination of WikiNEuRal with other datasets (e.g. WikiNEuRal + CoNLL). ## Licensing Information Contents of this repository are restricted to only non-commercial research purposes under the [Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (CC BY-NC-SA 4.0)](https://creativecommons.org/licenses/by-nc-sa/4.0/). Copyright of the dataset contents and models belongs to the original copyright holders.
{"language": ["de", "en", "es", "fr", "it", "nl", "pl", "pt", "ru", "multilingual"], "license": ["cc-by-nc-sa-4.0"], "tags": ["named-entity-recognition", "sequence-tagger-model"], "datasets": ["Babelscape/wikineural"], "annotations_creators": ["machine-generated"], "language_creators": ["machine-generated"], "widget": [{"text": "My name is Wolfgang and I live in Berlin."}, {"text": "George Washington went to Washington."}, {"text": "Mi nombre es Sarah y vivo en Londres."}, {"text": "\u041c\u0435\u043d\u044f \u0437\u043e\u0432\u0443\u0442 \u0421\u0438\u043c\u043e\u043d\u0430, \u0438 \u044f \u0436\u0438\u0432\u0443 \u0432 \u0420\u0438\u043c\u0435."}], "pretty_name": "wikineural-dataset", "source_datasets": ["original"], "task_categories": ["structure-prediction"], "task_ids": ["named-entity-recognition"]}
token-classification
Babelscape/wikineural-multilingual-ner
[ "transformers", "pytorch", "tensorboard", "safetensors", "bert", "token-classification", "named-entity-recognition", "sequence-tagger-model", "de", "en", "es", "fr", "it", "nl", "pl", "pt", "ru", "multilingual", "dataset:Babelscape/wikineural", "license:cc-by-nc-sa-4.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[ "de", "en", "es", "fr", "it", "nl", "pl", "pt", "ru", "multilingual" ]
TAGS #transformers #pytorch #tensorboard #safetensors #bert #token-classification #named-entity-recognition #sequence-tagger-model #de #en #es #fr #it #nl #pl #pt #ru #multilingual #dataset-Babelscape/wikineural #license-cc-by-nc-sa-4.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# WikiNEuRal: Combined Neural and Knowledge-based Silver Data Creation for Multilingual NER This is the model card for the EMNLP 2021 paper WikiNEuRal: Combined Neural and Knowledge-based Silver Data Creation for Multilingual NER. We fine-tuned a multilingual language model (mBERT) for 3 epochs on our WikiNEuRal dataset for Named Entity Recognition (NER). The resulting multilingual NER model supports the 9 languages covered by WikiNEuRal (de, en, es, fr, it, nl, pl, pt, ru), and it was trained on all 9 languages jointly. If you use the model, please reference this work in your paper: The original repository for the paper can be found at URL ## How to use You can use this model with Transformers *pipeline* for NER. ## Limitations and bias This model is trained on WikiNEuRal, a state-of-the-art dataset for Multilingual NER automatically derived from Wikipedia. Therefore, it might not generalize well to all textual genres (e.g. news). On the other hand, models trained only on news articles (e.g. only on CoNLL03) have been proven to obtain much lower scores on encyclopedic articles. To obtain more robust systems, we encourage you to train a system on the combination of WikiNEuRal with other datasets (e.g. WikiNEuRal + CoNLL). ## Licensing Information Contents of this repository are restricted to only non-commercial research purposes under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (CC BY-NC-SA 4.0). Copyright of the dataset contents and models belongs to the original copyright holders.
[ "# WikiNEuRal: Combined Neural and Knowledge-based Silver Data Creation for Multilingual NER\nThis is the model card for the EMNLP 2021 paper WikiNEuRal: Combined Neural and Knowledge-based Silver Data Creation for Multilingual NER. We fine-tuned a multilingual language model (mBERT) for 3 epochs on our WikiNEuRal dataset for Named Entity Recognition (NER). The resulting multilingual NER model supports the 9 languages covered by WikiNEuRal (de, en, es, fr, it, nl, pl, pt, ru), and it was trained on all 9 languages jointly.\n\nIf you use the model, please reference this work in your paper:\n\n\n \nThe original repository for the paper can be found at URL", "## How to use\n\nYou can use this model with Transformers *pipeline* for NER.", "## Limitations and bias\n\nThis model is trained on WikiNEuRal, a state-of-the-art dataset for Multilingual NER automatically derived from Wikipedia. Therefore, it might not generalize well to all textual genres (e.g. news). On the other hand, models trained only on news articles (e.g. only on CoNLL03) have been proven to obtain much lower scores on encyclopedic articles. To obtain more robust systems, we encourage you to train a system on the combination of WikiNEuRal with other datasets (e.g. WikiNEuRal + CoNLL).", "## Licensing Information\n\nContents of this repository are restricted to only non-commercial research purposes under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (CC BY-NC-SA 4.0). Copyright of the dataset contents and models belongs to the original copyright holders." ]
[ "TAGS\n#transformers #pytorch #tensorboard #safetensors #bert #token-classification #named-entity-recognition #sequence-tagger-model #de #en #es #fr #it #nl #pl #pt #ru #multilingual #dataset-Babelscape/wikineural #license-cc-by-nc-sa-4.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# WikiNEuRal: Combined Neural and Knowledge-based Silver Data Creation for Multilingual NER\nThis is the model card for the EMNLP 2021 paper WikiNEuRal: Combined Neural and Knowledge-based Silver Data Creation for Multilingual NER. We fine-tuned a multilingual language model (mBERT) for 3 epochs on our WikiNEuRal dataset for Named Entity Recognition (NER). The resulting multilingual NER model supports the 9 languages covered by WikiNEuRal (de, en, es, fr, it, nl, pl, pt, ru), and it was trained on all 9 languages jointly.\n\nIf you use the model, please reference this work in your paper:\n\n\n \nThe original repository for the paper can be found at URL", "## How to use\n\nYou can use this model with Transformers *pipeline* for NER.", "## Limitations and bias\n\nThis model is trained on WikiNEuRal, a state-of-the-art dataset for Multilingual NER automatically derived from Wikipedia. Therefore, it might not generalize well to all textual genres (e.g. news). On the other hand, models trained only on news articles (e.g. only on CoNLL03) have been proven to obtain much lower scores on encyclopedic articles. To obtain more robust systems, we encourage you to train a system on the combination of WikiNEuRal with other datasets (e.g. WikiNEuRal + CoNLL).", "## Licensing Information\n\nContents of this repository are restricted to only non-commercial research purposes under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (CC BY-NC-SA 4.0). Copyright of the dataset contents and models belongs to the original copyright holders." ]
[ 115, 186, 22, 142, 68 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #safetensors #bert #token-classification #named-entity-recognition #sequence-tagger-model #de #en #es #fr #it #nl #pl #pt #ru #multilingual #dataset-Babelscape/wikineural #license-cc-by-nc-sa-4.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# WikiNEuRal: Combined Neural and Knowledge-based Silver Data Creation for Multilingual NER\nThis is the model card for the EMNLP 2021 paper WikiNEuRal: Combined Neural and Knowledge-based Silver Data Creation for Multilingual NER. We fine-tuned a multilingual language model (mBERT) for 3 epochs on our WikiNEuRal dataset for Named Entity Recognition (NER). The resulting multilingual NER model supports the 9 languages covered by WikiNEuRal (de, en, es, fr, it, nl, pl, pt, ru), and it was trained on all 9 languages jointly.\n\nIf you use the model, please reference this work in your paper:\n\n\n \nThe original repository for the paper can be found at URL## How to use\n\nYou can use this model with Transformers *pipeline* for NER.## Limitations and bias\n\nThis model is trained on WikiNEuRal, a state-of-the-art dataset for Multilingual NER automatically derived from Wikipedia. Therefore, it might not generalize well to all textual genres (e.g. news). On the other hand, models trained only on news articles (e.g. only on CoNLL03) have been proven to obtain much lower scores on encyclopedic articles. To obtain more robust systems, we encourage you to train a system on the combination of WikiNEuRal with other datasets (e.g. WikiNEuRal + CoNLL)." ]
[ -0.03533551096916199, 0.15196262300014496, -0.0057596866972744465, 0.043803174048662186, 0.02020713873207569, -0.005791347939521074, 0.1037941500544548, 0.09248290210962296, -0.04353031888604164, 0.08420427143573761, 0.03668319061398506, 0.07187390327453613, 0.06760594248771667, 0.004351846408098936, 0.0635075643658638, -0.24858811497688293, 0.04668787866830826, -0.03304719924926758, 0.15894219279289246, 0.057031601667404175, 0.08319365233182907, -0.03605473041534424, 0.031512994319200516, 0.026157738640904427, -0.07507186383008957, 0.031043054535984993, -0.044877439737319946, -0.04433755576610565, 0.1028287336230278, 0.06806915998458862, 0.1014530137181282, 0.021802173927426338, 0.051235854625701904, -0.18711557984352112, 0.02627260610461235, 0.07780202478170395, -0.010744551196694374, 0.0153273930773139, 0.06701736897230148, -0.07057604938745499, 0.1540994495153427, -0.08396249264478683, 0.041622381657361984, 0.023450106382369995, -0.08778449147939682, -0.14910918474197388, -0.13613077998161316, 0.11617275327444077, 0.005547477398067713, 0.05125221610069275, -0.021780425682663918, 0.1344623863697052, 0.03389184549450874, 0.048181287944316864, 0.09007305651903152, -0.2295198291540146, -0.05163818225264549, 0.13392069935798645, 0.0036526054609566927, -0.023456906899809837, -0.05025644972920418, 0.02835484780371189, 0.04570899158716202, 0.0665624588727951, -0.03235901519656181, -0.02131732925772667, 0.036599867045879364, -0.04045271873474121, -0.13866053521633148, -0.0056099314242601395, 0.04431167244911194, -0.005013504531234503, -0.10395947098731995, -0.10275045037269592, -0.03866490721702576, 0.1715986430644989, 0.04344026371836662, -0.018857279792428017, -0.0035843919031322002, 0.005337180104106665, 0.07665304839611053, -0.13740114867687225, -0.08865498751401901, 0.029583381488919258, -0.102745421230793, 0.1935262233018875, 0.06794033199548721, 0.033588748425245285, 0.027742229402065277, 0.10189501941204071, -0.10955224186182022, -0.029405508190393448, -0.05637231469154358, -0.08260001987218857, -0.05273311212658882, -0.012874696403741837, -0.02543465979397297, -0.14013275504112244, 0.03687727451324463, 0.13913467526435852, -0.0520792230963707, 0.024424679577350616, -0.04367051273584366, 0.031724173575639725, 0.08697464317083359, 0.09740651398897171, -0.08023180067539215, -0.001467113965190947, 0.02268805541098118, -0.031721081584692, 0.06618763506412506, -0.016157260164618492, -0.03295464441180229, 0.003878328250721097, -0.008005920797586441, 0.03728862479329109, 0.02237667888402939, 0.025645961984992027, 0.0010997045319527388, 0.002053419826552272, 0.13373352587223053, -0.14487776160240173, 0.02609521895647049, -0.02680068649351597, -0.009558873251080513, 0.2585843503475189, 0.031966231763362885, -0.012138749472796917, -0.07566900551319122, 0.07739928364753723, -0.013555971905589104, 0.01958942413330078, -0.06921637803316116, -0.09548880904912949, 0.0713016539812088, -0.09189370274543762, -0.03727376088500023, -0.11031757295131683, -0.13931511342525482, -0.0772441029548645, 0.02455582097172737, -0.03836827725172043, 0.009999924339354038, -0.03059048391878605, 0.025773337110877037, -0.0036462116986513138, 0.026465630158782005, -0.07010966539382935, -0.010277519933879375, -0.015010337345302105, -0.09897521138191223, 0.03010629117488861, -0.08154784142971039, -0.012027764692902565, -0.049876268953084946, 0.03146138787269592, -0.1589636653661728, 0.07131797820329666, -0.07806938886642456, 0.009838799014687538, -0.09033411741256714, -0.055833850055933, -0.045113589614629745, 0.040646251291036606, 0.03268851712346077, 0.04245668277144432, -0.18016716837882996, -0.02998817339539528, 0.1955772340297699, -0.18752235174179077, 0.030064783990383148, 0.11191902309656143, -0.021261470392346382, 0.03582907095551491, 0.10204371809959412, 0.0918600782752037, 0.09164499491453171, -0.08303207904100418, -0.12453637272119522, -0.03940689191222191, -0.02915208227932453, 0.12047211080789566, 0.017562827095389366, -0.04470840096473694, 0.022247683256864548, 0.01761864870786667, 0.016735609620809555, -0.04518644139170647, 0.003111095866188407, -0.00783488992601633, -0.02888377755880356, -0.009276935830712318, 0.002449962543323636, -0.02007358893752098, -0.019059734418988228, -0.030408333986997604, -0.10039404779672623, 0.12236060947179794, 0.11347214132547379, -0.06682789325714111, 0.06587600708007812, -0.07274501025676727, 0.09128741174936295, -0.024276217445731163, 0.01370937004685402, -0.14475466310977936, -0.10207716375589371, 0.037763360887765884, -0.16643382608890533, 0.10798883438110352, 0.015668120235204697, 0.017180925235152245, 0.07177770137786865, -0.027488388121128082, 0.010370142757892609, -0.04479626566171646, -0.01863807998597622, 0.0019801275338977575, -0.08952007442712784, -0.03413018211722374, -0.07064250856637955, 0.10125759989023209, -0.1368446946144104, 0.022594135254621506, 0.06178366392850876, 0.13297328352928162, -0.0051214131526649, -0.021789494901895523, 0.028392955660820007, 0.054149895906448364, -0.0257707629352808, -0.03663686290383339, 0.03265879303216934, 0.0032851325813680887, -0.06947652995586395, 0.06365789473056793, -0.05992359295487404, -0.17360374331474304, 0.06698345392942429, 0.04572228342294693, -0.07901085913181305, 0.009084112010896206, -0.013001391664147377, -0.03220561519265175, -0.10292026400566101, -0.03248457983136177, 0.1856936663389206, 0.0481293648481369, 0.032819684594869614, -0.08275823295116425, -0.0417759008705616, -0.012821480631828308, -0.04760991781949997, -0.04523434862494469, 0.06663259118795395, 0.018214348703622818, -0.16607503592967987, 0.03078347072005272, 0.01912131905555725, -0.038351088762283325, 0.12994825839996338, 0.03056596778333187, -0.05914461240172386, -0.00622200220823288, 0.002545083174481988, 0.028633665293455124, 0.019918235018849373, 0.047165412455797195, 0.011025040410459042, 0.009335759095847607, 0.02901671826839447, 0.030945226550102234, -0.09739217162132263, 0.050387267023324966, -0.011636951006948948, -0.01522965170443058, 0.03152264282107353, 0.04044698551297188, 0.0002123200974892825, 0.06201590597629547, -0.018481573089957237, 0.05650835856795311, -0.03459934890270233, -0.010635276325047016, -0.06704382598400116, 0.1533891260623932, -0.10417605936527252, -0.1980883628129959, -0.14896860718727112, 0.0031223143450915813, -0.0919656828045845, -0.06263501942157745, 0.023303236812353134, -0.013552202843129635, -0.05229155719280243, -0.11771459132432938, 0.046573907136917114, -0.02955400012433529, -0.0717194527387619, -0.03426849842071533, -0.013236266560852528, -0.021238898858428, -0.14443863928318024, -0.0027769978623837233, -0.0020570161286741495, -0.09059508889913559, 0.03779922053217888, 0.01817559450864792, 0.03880278021097183, 0.09798679500818253, 0.015388146974146366, -0.015547657385468483, -0.008032703772187233, 0.10494150966405869, -0.05449909716844559, 0.1358039230108261, 0.11291688680648804, -0.040807031095027924, 0.08581484854221344, 0.041206952184438705, 0.056002963334321976, -0.05579066649079323, 0.006085613276809454, 0.08060826361179352, -0.05772794410586357, -0.25990933179855347, -0.06512372940778732, 0.008571505546569824, -0.004564744886010885, 0.016406243667006493, 0.0683862715959549, -0.05488419905304909, 0.03748297318816185, -0.030792122706770897, -0.04231429472565651, 0.037784479558467865, 0.07209846377372742, 0.1301264762878418, -0.01242106407880783, 0.07125414907932281, -0.08153465390205383, -0.0022344947792589664, 0.10571760684251785, 0.04855077713727951, 0.18005692958831787, 0.0005916169611737132, -0.004811674356460571, 0.10369331389665604, 0.022033123299479485, 0.04654321447014809, 0.08808101713657379, -0.04153986647725105, 0.03341066464781761, -0.008861836977303028, -0.061260491609573364, 0.0709800124168396, 0.06183537468314171, 0.020744798704981804, -0.021613677963614464, 0.053903382271528244, -0.06778362393379211, 0.06386786699295044, 0.17844714224338531, 0.08157069981098175, -0.12451904267072678, -0.0359518937766552, 0.03375094756484032, -0.07800444960594177, -0.04638219252228737, 0.006256913300603628, 0.024181677028536797, -0.15582667291164398, 0.15252384543418884, -0.0014982273569330573, 0.056980401277542114, -0.18143019080162048, -0.0534726046025753, -0.043388571590185165, 0.08575829118490219, -0.023382993414998055, 0.07235785573720932, -0.08562839776277542, 0.09917516261339188, -0.005172614008188248, 0.06208055093884468, -0.01218294259160757, 0.03564830496907234, 0.028477024286985397, 0.07468181103467941, 0.10576172173023224, 0.02739754132926464, -0.07615849375724792, 0.027665099129080772, -0.12497090548276901, 0.0007595791830681264, 0.06863381713628769, -0.020234577357769012, 0.10358735173940659, 0.03832324966788292, 0.003641492919996381, -0.08134907484054565, 0.0036114598624408245, -0.16059362888336182, -0.1818205863237381, 0.03037526085972786, -0.06313741952180862, -0.00047416024608537555, -0.03539547324180603, -0.09872154891490936, -0.12341032177209854, 0.1658668965101242, -0.043808311223983765, -0.08570537716150284, -0.09432915598154068, -0.019589275121688843, 0.09517192095518112, -0.054711099714040756, 0.019932493567466736, -0.026959741488099098, 0.09734103083610535, -0.1111239343881607, -0.06679702550172806, -0.029351606965065002, -0.039395418018102646, -0.1089760959148407, 0.019226256757974625, 0.10891763120889664, 0.08863828331232071, 0.007763299159705639, 0.034435439854860306, 0.045028842985630035, 0.03098824806511402, -0.1445952206850052, 0.016743401065468788, 0.19599243998527527, 0.019551729783415794, 0.06033128499984741, -0.04750371351838112, -0.06672032922506332, -0.0744052603840828, -0.05013642832636833, 0.13962852954864502, 0.1859510987997055, -0.06229232996702194, 0.08510388433933258, 0.1387416422367096, -0.135577991604805, -0.19723378121852875, -0.03956405445933342, 0.008486342616379261, 0.014682289212942123, -0.04994310438632965, -0.1908881962299347, 0.06829911470413208, 0.1251985728740692, 0.012923267669975758, -0.0033480278216302395, -0.207911878824234, -0.11522150039672852, 0.017147846519947052, 0.0480903722345829, 0.047444555908441544, -0.08042234182357788, -0.05058949440717697, -0.032828886061906815, -0.04953324794769287, 0.1033545583486557, -0.1407688409090042, 0.04765785485506058, -0.0013217957457527518, 0.04794108122587204, 0.0077938358299434185, -0.025715550407767296, 0.11286045610904694, 0.019804900512099266, 0.017562933266162872, -0.04320776090025902, 0.06309833377599716, 0.024047335609793663, -0.020441006869077682, 0.15629920363426208, 0.04721894487738609, -0.010088032111525536, -0.044059690088033676, -0.06905854493379593, -0.053786080330610275, 0.08886467665433884, -0.060282159596681595, -0.03899066150188446, -0.05139923840761185, 0.06479059159755707, -0.00021238236513454467, -0.020347291603684425, 0.011530236341059208, -0.05867190286517143, 0.06551355123519897, 0.08396682888269424, 0.11687415838241577, 0.04166262596845627, -0.04950689896941185, -0.010209414176642895, -0.05214306712150574, 0.02326560579240322, -0.0741942748427391, -0.021227654069662094, 0.11130601167678833, 0.015856314450502396, 0.08407498896121979, 0.0042643556371331215, -0.20639154314994812, 0.016152869910001755, 0.058825984597206116, -0.11655675619840622, -0.15360717475414276, 0.022856928408145905, 0.018745802342891693, 0.022590110078454018, 0.03643083572387695, 0.1836284101009369, -0.06660861521959305, -0.03937704488635063, 0.022495953366160393, 0.04207829758524895, -0.018129998818039894, 0.0287312064319849, 0.027034593746066093, -0.0055532436817884445, -0.03413520008325577, 0.1272301822900772, 0.10123393684625626, -0.12113853543996811, 0.006276159081608057, 0.11762876063585281, -0.057908304035663605, -0.07539842277765274, -0.09394953399896622, 0.047834668308496475, -0.11958910524845123, -0.10402087867259979, 0.032333854585886, -0.10436862707138062, 0.00579509511590004, 0.15569449961185455, 0.030618639662861824, 0.05211995914578438, -0.007307542487978935, -0.04556769132614136, -0.03905128315091133, 0.02245718613266945, 0.04174629598855972, 0.03554944694042206, -0.049202289432287216, 0.029683677479624748, 0.05341077595949173, 0.02331199310719967, 0.003979390021413565, -0.04384379833936691, -0.07623391598463058, -0.022145865485072136, -0.0880233645439148, 0.027643561363220215, -0.10370872169733047, 0.013504973612725735, 0.011492729187011719, 0.002607665490359068, 0.035554200410842896, -0.00722184544429183, -0.03504510968923569, -0.016398832201957703, -0.0229378342628479, 0.07867764681577682, -0.11218839883804321, 0.004233292303979397, 0.031195018440485, -0.06637337058782578, 0.08275732398033142, 0.0030440203845500946, -0.0195161122828722, 0.08209115266799927, -0.06737552583217621, 0.055835772305727005, -0.00006104161002440378, 0.028349243104457855, -0.011496610939502716, -0.12174977362155914, -0.002574996091425419, 0.00548614701256156, -0.03866158425807953, 0.031081311404705048, 0.00899500586092472, -0.059039175510406494, 0.05929073691368103, 0.04513285681605339, -0.05503565073013306, -0.03884495794773102, 0.01949419267475605, 0.04473653435707092, 0.07217317074537277, 0.07388937473297119, -0.07422109693288803, 0.056574732065200806, -0.07694463431835175, -0.002043876564130187, 0.007060590665787458, -0.019953591749072075, 0.06960972398519516, -0.026788417249917984, 0.028640419244766235, 0.02036462351679802, 0.16680343449115753, -0.06778427213430405, 0.012868337333202362, 0.010183587670326233, -0.06563087552785873, -0.16021300852298737, 0.010309397242963314, -0.03414006531238556, 0.0039201597683131695, 0.011569707654416561, -0.07794656604528427, -0.05257691815495491, -0.062337476760149, -0.036452509462833405, 0.11985553801059723, 0.07698700577020645, 0.13909581303596497, 0.06749194860458374, 0.07705134153366089, 0.03698083013296127, -0.04199843853712082, 0.017097603529691696, -0.003283495083451271, 0.036851659417152405, -0.05019643157720566, 0.09235011041164398, 0.1221374124288559, -0.10302439332008362, 0.0774589329957962, 0.05709446966648102, -0.052894171327352524, -0.11236783117055893, -0.22934529185295105, -0.04901452362537384, 0.004303842782974243, 0.01395002193748951, -0.0954461544752121, 0.0672965794801712, 0.03093530982732773, 0.0679892748594284, -0.007182328030467033, 0.058024995028972626, -0.03489227592945099, -0.06547611951828003, 0.06272261589765549, 0.02138635516166687, 0.0227859765291214, 0.05203434079885483, 0.048486754298210144, -0.04637858644127846, 0.06183571740984917, 0.03823360055685043, 0.0918259397149086, 0.09930919855833054, 0.031400613486766815, -0.051062121987342834, -0.0248074010014534, -0.009900232776999474, 0.0024498829152435064, 0.023141534999012947, 0.05574716627597809, 0.09075374901294708, -0.06256113946437836, -0.011127367615699768, 0.10309480875730515, -0.02252797782421112, -0.11021653562784195, -0.12116572260856628, 0.193729966878891, -0.022754903882741928, 0.022323599085211754, 0.01663218066096306, -0.08486751466989517, 0.02347351610660553, 0.12432350963354111, 0.13909409940242767, 0.06452296674251556, 0.025278698652982712, 0.016136769205331802, 0.006485105026513338, 0.009141865186393261, 0.039612773805856705, 0.02188611775636673, 0.2017950564622879, -0.06184626370668411, 0.03083009086549282, -0.10034959763288498, 0.0074301352724432945, -0.09017758816480637, 0.07896094024181366, -0.05703195184469223, -0.02522827684879303, -0.04415736719965935, 0.11287294328212738, -0.10425271838903427, -0.20964780449867249, 0.010558146983385086, -0.03239724040031433, -0.10613253712654114, -0.003934302367269993, 0.0015802161069586873, -0.025602741166949272, 0.06378781050443649, -0.004231590311974287, -0.03386475518345833, 0.1547965556383133, 0.01331701222807169, -0.04985099658370018, -0.02562662959098816, 0.05706489458680153, -0.006435462273657322, 0.08589784055948257, 0.028789116069674492, 0.07275078445672989, 0.08460496366024017, 0.010869591496884823, -0.07714714109897614, 0.08067697286605835, -0.006458333693444729, -0.035359229892492294, -0.013093781657516956, 0.11229534447193146, 0.022797370329499245, 0.06880401819944382, 0.047097619622945786, -0.013111804611980915, 0.07619249820709229, 0.1165112555027008, -0.03448745980858803, -0.054554373025894165, 0.12495958805084229, -0.12170272320508957, 0.1615486890077591, 0.15992756187915802, -0.004802534822374582, 0.0006544881034642458, -0.02979758381843567, 0.022594738751649857, -0.04034493491053581, 0.02905960939824581, -0.0006707696593366563, -0.10664359480142593, 0.04474176466464996, -0.11817725002765656, 0.09328774362802505, -0.16774919629096985, -0.06559622287750244, 0.056975822895765305, -0.008282521739602089, -0.06244959682226181, 0.06867013871669769, 0.05594484135508537, 0.030059952288866043, -0.03434773534536362, -0.09684529900550842, 0.002309835981577635, 0.05066889524459839, -0.0804283618927002, -0.0466681644320488 ]
null
null
transformers
#Anika Bot
{"tags": ["conversational"]}
text-generation
Backedman/DialoGPT-small-Anika
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
#Anika Bot
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.009697278961539268, 0.03208012506365776, -0.007204889785498381, 0.004809224978089333, 0.16726240515708923, 0.014898733235895634, 0.09765533357858658, 0.13672804832458496, -0.007841327227652073, -0.031050153076648712, 0.14490588009357452, 0.20411323010921478, -0.006439372431486845, 0.0661218985915184, -0.07572533935308456, -0.2683109939098358, 0.05759621039032936, 0.046649303287267685, 0.016515716910362244, 0.1200079694390297, 0.08573378622531891, -0.05473608896136284, 0.08714032918214798, -0.014583407901227474, -0.150366872549057, 0.017733458429574966, 0.043394338339567184, -0.12260226160287857, 0.11910516023635864, 0.05462685227394104, 0.07063519209623337, 0.014929565601050854, -0.07541623711585999, -0.1631229966878891, 0.03031250834465027, 0.01425902172923088, -0.0594632662832737, 0.04757995903491974, 0.059961482882499695, -0.10165371745824814, 0.10819483548402786, 0.09530027210712433, -0.013078106567263603, 0.06798283755779266, -0.16849711537361145, -0.020869607105851173, -0.01446688175201416, 0.009899779222905636, 0.05550243332982063, 0.09964893013238907, -0.03413357585668564, 0.10497362166643143, -0.09214533120393753, 0.11017382889986038, 0.10932035744190216, -0.32057443261146545, -0.005767723545432091, 0.09167823940515518, 0.039358653128147125, 0.07352814823389053, -0.04467793554067612, 0.06258884817361832, 0.018015462905168533, 0.017986174672842026, -0.014015024527907372, -0.07283061742782593, -0.11612214148044586, 0.04717336222529411, -0.08668071031570435, -0.059868961572647095, 0.2244078367948532, -0.05464440956711769, 0.06881742179393768, -0.05281897634267807, -0.10522868484258652, -0.04308144748210907, -0.029833965003490448, 0.00475557055324316, -0.07660607248544693, 0.08692064881324768, 0.00869679357856512, -0.09547875821590424, -0.1376667022705078, -0.02496783249080181, -0.1776352822780609, 0.16140350699424744, 0.02465328387916088, 0.05232657864689827, -0.2027255892753601, 0.09623090922832489, 0.017906051129102707, -0.08045592904090881, 0.022091427817940712, -0.10046248883008957, 0.029131146147847176, 0.013760408386588097, -0.04754498973488808, -0.061387211084365845, 0.0843690037727356, 0.11199145019054413, -0.01731434464454651, 0.025486016646027565, -0.039331406354904175, 0.08100687712430954, 0.03553595021367073, 0.09077847748994827, 0.007288969587534666, -0.028338588774204254, 0.025842782109975815, -0.13719046115875244, -0.003647835226729512, -0.07116208970546722, -0.16572439670562744, -0.021088803187012672, 0.02994808368384838, 0.08289173990488052, 0.015449047088623047, 0.11682453751564026, -0.03272046521306038, -0.025152435526251793, 0.03602350503206253, -0.047656361013650894, -0.012649794109165668, 0.016648368909955025, 0.013163427822291851, 0.12399329990148544, -0.0022096503525972366, 0.03235051408410072, -0.13653022050857544, 0.031423524022102356, -0.06793295592069626, -0.003740974934771657, -0.03486552834510803, -0.040637075901031494, 0.009043924510478973, -0.06862333416938782, 0.003486064961180091, -0.15030112862586975, -0.15063877403736115, 0.007587034720927477, -0.007836631499230862, -0.04107699543237686, -0.06370922178030014, -0.06952770054340363, -0.013550350442528725, 0.04251532256603241, -0.07093454152345657, -0.011352915316820145, -0.06403283774852753, 0.11004766076803207, -0.03197755664587021, 0.07921615242958069, -0.11953279376029968, 0.08390819281339645, -0.11260783672332764, -0.02386913076043129, -0.060801517218351364, 0.09317506104707718, -0.0006014376995153725, 0.09549830108880997, -0.006563255097717047, -0.017931854352355003, -0.07981178909540176, 0.06445012241601944, -0.042872510850429535, 0.21701598167419434, -0.0615808479487896, -0.11181682348251343, 0.28781595826148987, -0.052628401666879654, -0.1370542049407959, 0.11647392809391022, 0.008682746440172195, 0.05777018144726753, 0.10703510791063309, 0.19733482599258423, -0.015276194550096989, 0.004040541127324104, 0.09471915662288666, 0.11263324320316315, -0.11276852339506149, -0.033160366117954254, 0.013019153848290443, -0.04081077128648758, -0.10867965966463089, 0.04689536616206169, 0.09810488671064377, 0.07090286910533905, -0.04786505550146103, -0.03377414867281914, -0.01366397924721241, 0.0052589005790650845, 0.08885077387094498, -0.007157256826758385, 0.10962837189435959, -0.05819983780384064, -0.03796621412038803, -0.029282379895448685, -0.012126247398555279, -0.03951939567923546, 0.03137664496898651, -0.043376367539167404, 0.10821941494941711, -0.011204327456653118, 0.06364280730485916, -0.16185984015464783, -0.07691477984189987, -0.017002692446112633, 0.1581239402294159, 0.024538565427064896, 0.09859629720449448, 0.0552486926317215, -0.040398042649030685, -0.0012767292791977525, 0.012792680412530899, 0.15581141412258148, -0.022091681137681007, -0.065607450902462, -0.052166227251291275, 0.08642971515655518, -0.05641226842999458, 0.04504093527793884, -0.05937713757157326, 0.012367865070700645, 0.05064384639263153, 0.10342344641685486, -0.00018274025933351368, 0.03323284164071083, -0.008164864964783192, 0.002145637758076191, -0.058205123990774155, 0.007405933458358049, 0.10799351334571838, 0.00036868182360194623, -0.07365862280130386, 0.22074243426322937, -0.17796069383621216, 0.1765957772731781, 0.1893044263124466, -0.299345999956131, 0.017949223518371582, -0.10759581625461578, -0.04561871662735939, 0.014407722279429436, 0.05567655712366104, -0.0454222597181797, 0.1703362911939621, -0.009871348738670349, 0.18874616920948029, -0.04946064203977585, -0.04464937001466751, -0.0200483538210392, -0.05118836089968681, -0.0024189651012420654, 0.07781197130680084, 0.10685696452856064, -0.13992026448249817, 0.1964332014322281, 0.1621224284172058, 0.048237916082143784, 0.19945049285888672, 0.015346456319093704, -0.011589210480451584, 0.0909530371427536, 0.005220826715230942, -0.058739423751831055, -0.07409929484128952, -0.2594851851463318, -0.030033592134714127, 0.07992640137672424, 0.0422382652759552, 0.1212305948138237, -0.11349532753229141, -0.038956157863140106, -0.01763172075152397, -0.023146281018853188, 0.021672505885362625, 0.0914369598031044, 0.06075398623943329, 0.13201528787612915, -0.001710098935291171, -0.007300339173525572, 0.10524573177099228, 0.01783694699406624, -0.09354141354560852, 0.18308524787425995, -0.13652534782886505, -0.37097251415252686, -0.13911493122577667, -0.18057456612586975, -0.05449081212282181, 0.05712554603815079, 0.11679314076900482, -0.12011238187551498, -0.018752124160528183, 0.01578843593597412, 0.10931742936372757, -0.08449502289295197, 0.0021454424131661654, -0.06880278885364532, 0.0321490578353405, -0.10310184955596924, -0.09194442629814148, -0.055416494607925415, -0.031392451375722885, -0.08001253753900528, 0.1423761546611786, -0.10777941346168518, 0.04476889222860336, 0.20262959599494934, 0.04653622955083847, 0.05625178664922714, -0.044105201959609985, 0.19377262890338898, -0.11264272034168243, -0.01661740615963936, 0.19215328991413116, -0.048360925167798996, 0.07476246356964111, 0.1232115849852562, -0.006348740309476852, -0.08765771239995956, 0.03011748194694519, -0.02085109055042267, -0.07988511025905609, -0.23219464719295502, -0.13938382267951965, -0.12429051846265793, 0.09477275609970093, 0.028005298227071762, 0.056365787982940674, 0.17219258844852448, 0.06577219814062119, -0.038416244089603424, 0.006410336587578058, 0.02959546446800232, 0.08237514644861221, 0.23417828977108002, -0.06035616248846054, 0.1364797055721283, -0.03420931473374367, -0.14982740581035614, 0.08169995993375778, 0.0713929831981659, 0.10213395953178406, 0.06678459793329239, 0.0804823637008667, 0.0149586396291852, 0.06188136339187622, 0.1311223804950714, 0.08191446959972382, 0.019586285576224327, -0.02480296604335308, -0.03388110175728798, -0.025523077696561813, -0.05937909707427025, 0.040128443390131, 0.06589099019765854, -0.16763372719287872, -0.039227183908224106, -0.09338314831256866, 0.09657008945941925, 0.0873042419552803, 0.06609832495450974, -0.1842060089111328, -0.008006223477423191, 0.08488986641168594, -0.03854905813932419, -0.13727426528930664, 0.09535189718008041, 0.01523482333868742, -0.15144726634025574, 0.03139317408204079, -0.04061909019947052, 0.12188644707202911, -0.07804752141237259, 0.09809603542089462, -0.08108244836330414, -0.07448557764291763, 0.02123199962079525, 0.1261177361011505, -0.30527687072753906, 0.20240111649036407, -0.0024993624538183212, -0.06486981362104416, -0.1243603527545929, -0.0032166161108762026, 0.002410882618278265, 0.07357452809810638, 0.10519039630889893, -0.007196315098553896, 0.001897757756523788, -0.06300821900367737, -0.01829923689365387, 0.032471053302288055, 0.13080233335494995, -0.0401318334043026, -0.021158374845981598, -0.050194524228572845, -0.001653497340157628, -0.03173094615340233, -0.06934895366430283, 0.02002747356891632, -0.19509181380271912, 0.08751901984214783, 0.04166261479258537, 0.09648149460554123, 0.029994789510965347, 0.004265148192644119, -0.09651939570903778, 0.24698667228221893, -0.07148019969463348, -0.10072879493236542, -0.10919588059186935, -0.046813901513814926, 0.03569883480668068, -0.05628936365246773, 0.04309194162487984, -0.0788632407784462, 0.028997479006648064, -0.06352769583463669, -0.19235502183437347, 0.12410202622413635, -0.09027006477117538, -0.04412810131907463, -0.02371402643620968, 0.2110891044139862, -0.05598580464720726, 0.010335659608244896, 0.02930437959730625, 0.01208863127976656, -0.11645778268575668, -0.09678568691015244, 0.031018631532788277, -0.007351789623498917, 0.050603240728378296, 0.041841957718133926, -0.05915454775094986, -0.017138581722974777, -0.052199993282556534, -0.022926922887563705, 0.3496883809566498, 0.14231905341148376, -0.043836336582899094, 0.19347235560417175, 0.12347975373268127, -0.07452994585037231, -0.3159443140029907, -0.1066238060593605, -0.10937739163637161, -0.04680149629712105, -0.07012093812227249, -0.2002030611038208, 0.06474938243627548, 0.00662544509395957, -0.013415241613984108, 0.12749312818050385, -0.2561831772327423, -0.07571036368608475, 0.15906259417533875, -0.017980827018618584, 0.3745945692062378, -0.1168576180934906, -0.10926306992769241, -0.03950892388820648, -0.14175476133823395, 0.16968177258968353, -0.01989765651524067, 0.11221715062856674, -0.009765521623194218, 0.14388824999332428, 0.05548359826207161, -0.023479344323277473, 0.08544106781482697, 0.004999885335564613, -0.03290518373250961, -0.10304180532693863, -0.05676887184381485, 0.007092386484146118, 0.02477436140179634, 0.018026655539870262, -0.041834570467472076, 0.02227151393890381, -0.11731979995965958, -0.04657655209302902, -0.08982590585947037, 0.04431166127324104, 0.03899754583835602, -0.07325074821710587, -0.002380647463724017, -0.07165111601352692, -0.012272949330508709, 0.022334342822432518, 0.20356793701648712, -0.08029330521821976, 0.16448934376239777, 0.09239562600851059, 0.12419285625219345, -0.14376309514045715, -0.00019283240544609725, -0.0762530043721199, -0.05611240118741989, 0.07737895101308823, -0.09433035552501678, 0.058893077075481415, 0.10901971161365509, -0.04567738622426987, 0.08828683942556381, 0.10377411544322968, 0.008936077356338501, 0.003213887568563223, 0.10916902124881744, -0.2667325437068939, -0.0296600554138422, -0.07532413303852081, 0.000883326749317348, 0.09092561900615692, 0.08562852442264557, 0.18840822577476501, 0.025361526757478714, -0.04293036088347435, -0.002770674182102084, 0.028597986325621605, -0.039021048694849014, 0.051667019724845886, 0.001123449532315135, 0.01947369985282421, -0.1530752182006836, 0.072522833943367, 0.01490565575659275, -0.15215420722961426, 0.021316176280379295, 0.16572684049606323, -0.11656328290700912, -0.1283872276544571, -0.06520111113786697, 0.08313824236392975, -0.11755692958831787, -0.01578943058848381, -0.03279297426342964, -0.13145680725574493, 0.07992171496152878, 0.12629036605358124, 0.05557859688997269, 0.0972496047616005, -0.06061713397502899, -0.020469192415475845, -0.018721895292401314, -0.014099318534135818, -0.012384648434817791, -0.007667020428925753, -0.055978111922740936, 0.0590752474963665, -0.026677248999476433, 0.1425808072090149, -0.09221141785383224, -0.1037059873342514, -0.16142144799232483, 0.0374140702188015, -0.11013076454401016, -0.08825794607400894, -0.08821134269237518, -0.050188567489385605, 0.002360827289521694, -0.019856395199894905, -0.04037635400891304, -0.05829505994915962, -0.12300454825162888, 0.0338277705013752, -0.040771447122097015, 0.024727050215005875, -0.07512269169092178, 0.015856385231018066, 0.08507686108350754, -0.03285100311040878, 0.15655414760112762, 0.1450488418340683, -0.1006515845656395, 0.10741901397705078, -0.14806775748729706, -0.09138492494821548, 0.11116421222686768, 0.015329592861235142, 0.0449691042304039, 0.09723787009716034, 0.013362943194806576, 0.0635865181684494, 0.032776717096567154, 0.05308786407113075, 0.027619892731308937, -0.11959987878799438, 0.06483134627342224, -0.03626115620136261, -0.14700546860694885, -0.049338050186634064, -0.05282869189977646, 0.01647452637553215, 0.013054544106125832, 0.09622690081596375, -0.05301849544048309, 0.10698331147432327, -0.04055701196193695, 0.0346808135509491, 0.017554637044668198, -0.1730053424835205, -0.03816922754049301, -0.08538098633289337, 0.03681723028421402, 0.014741539023816586, 0.25266793370246887, 0.030072299763560295, 0.012416383251547813, 0.032671261578798294, 0.08285367488861084, 0.03899408504366875, 0.010228337720036507, 0.17482228577136993, 0.1162426546216011, -0.06621865928173065, -0.10445023328065872, 0.0729617029428482, 0.016332454979419708, 0.01286179106682539, 0.13617953658103943, 0.008365051820874214, 0.005795429926365614, 0.08649782836437225, -0.016865963116288185, 0.009968153201043606, -0.10052056610584259, -0.13426925241947174, -0.022176474332809448, 0.05151832848787308, -0.04655967652797699, 0.11727844923734665, 0.1406494379043579, -0.01806013658642769, 0.03222079202532768, -0.021771740168333054, -0.05699979141354561, -0.1683429479598999, -0.1429590880870819, -0.06883849948644638, -0.13416796922683716, 0.00897989235818386, -0.11180389672517776, 0.05395037308335304, 0.06001098081469536, 0.06750501692295074, -0.06899319589138031, 0.10220931470394135, 0.04626858979463577, -0.11440542340278625, 0.06264589726924896, -0.0296088308095932, 0.09430401772260666, -0.02759445086121559, -0.019505485892295837, -0.09039592742919922, 0.014574515633285046, 0.011419114656746387, 0.06245238706469536, -0.04707273095846176, 0.007463190704584122, -0.14696238934993744, -0.08972041308879852, -0.0523175448179245, 0.0718572810292244, -0.050409089773893356, 0.14282815158367157, 0.00775480642914772, -0.0170906875282526, 0.039554283022880554, 0.22787313163280487, -0.07476283609867096, -0.04778539761900902, -0.05269690603017807, 0.20717895030975342, 0.02975541539490223, 0.1171872541308403, -0.022938819602131844, -0.006106364540755749, -0.0919521227478981, 0.3764844834804535, 0.30030161142349243, -0.09031439572572708, 0.011794124729931355, 0.02137952297925949, 0.04502861574292183, 0.1316293478012085, 0.1216534823179245, 0.10318691283464432, 0.3006802201271057, -0.07452366501092911, -0.04653361067175865, -0.012629742734134197, -0.023858042433857918, -0.09059546142816544, 0.1021224707365036, 0.04839762672781944, -0.06382183730602264, -0.03313443064689636, 0.0954432487487793, -0.25862133502960205, 0.1277991235256195, -0.12311873584985733, -0.17578600347042084, -0.06654827296733856, 0.009760108776390553, 0.10465722531080246, 0.015642458572983742, 0.0946015790104866, 0.007128213066607714, -0.11252258718013763, 0.06305865943431854, 0.03397420793771744, -0.22762253880500793, 0.0006893770187161863, 0.06642123311758041, -0.07006710022687912, -0.0024247700348496437, -0.026499588042497635, 0.05657242611050606, 0.0656052976846695, 0.054629553109407425, -0.00971333310008049, 0.03816632181406021, 0.0034184439573436975, -0.0585215799510479, 0.016623929142951965, 0.05121519789099693, 0.02472509816288948, -0.09763528406620026, 0.06927435845136642, -0.1574270874261856, 0.04766253009438515, -0.0030655991286039352, -0.04124255105853081, 0.006064958870410919, 0.008823691867291927, -0.06491616368293762, 0.05165379121899605, 0.07916834205389023, -0.0016257909592241049, -0.0062433634884655476, -0.057178743183612823, -0.02632102556526661, -0.027755750343203545, -0.09291748702526093, -0.10495562851428986, -0.14682936668395996, -0.11640441417694092, 0.09368976950645447, -0.01011267676949501, -0.1848134547472, 0.022154374048113823, -0.08606051653623581, 0.08319322764873505, -0.1670055389404297, 0.08040720224380493, 0.07041648775339127, 0.013038921169936657, -0.0031511052511632442, -0.02002427540719509, 0.054132770746946335, 0.086809903383255, -0.10407156497240067, -0.07400695979595184 ]
null
null
transformers
Dataset used for training: - Name: Common Voice - Language: Indonesian [id] - Version: 6.1 Test WER: 19.3 % Contact: bagus@ep.its.ac.id
{"language": "el", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "bahasa-indonesia"], "datasets": ["common_voice_id_6.1"]}
automatic-speech-recognition
Bagus/wav2vec2-large-xlsr-bahasa-indonesia
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "bahasa-indonesia", "el", "dataset:common_voice_id_6.1", "license:apache-2.0", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[ "el" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #audio #speech #bahasa-indonesia #el #dataset-common_voice_id_6.1 #license-apache-2.0 #endpoints_compatible #has_space #region-us
Dataset used for training: - Name: Common Voice - Language: Indonesian [id] - Version: 6.1 Test WER: 19.3 % Contact: bagus@URL
[]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #audio #speech #bahasa-indonesia #el #dataset-common_voice_id_6.1 #license-apache-2.0 #endpoints_compatible #has_space #region-us \n" ]
[ 76 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #audio #speech #bahasa-indonesia #el #dataset-common_voice_id_6.1 #license-apache-2.0 #endpoints_compatible #has_space #region-us \n" ]
[ -0.06179199740290642, 0.0746750682592392, -0.0037977874744683504, -0.06584830582141876, 0.022364120930433273, -0.04548623040318489, 0.11895807832479477, 0.12703831493854523, 0.00395465362817049, 0.0004384141357149929, 0.059394992887973785, 0.08508629351854324, 0.07627295702695847, -0.01965235359966755, -0.023335972800850868, -0.23671483993530273, 0.12133226543664932, 0.00887489877641201, 0.1422702670097351, 0.10580432415008545, 0.09607625007629395, -0.03490501642227173, 0.03379148989915848, 0.06253094971179962, -0.0238770954310894, 0.014037346467375755, -0.008116716518998146, -0.17312033474445343, 0.09464620798826218, 0.00043058153823949397, 0.04449882730841637, 0.05371011421084404, -0.004154934082180262, -0.1833086758852005, 0.00604423601180315, -0.027703125029802322, 0.007440282963216305, -0.0039022001437842846, 0.002160636940971017, 0.030058961361646652, 0.03351166471838951, 0.08362259715795517, -0.03664621338248253, 0.04000101983547211, -0.04849172383546829, -0.28117865324020386, -0.07267196476459503, 0.05693682283163071, 0.06435328722000122, 0.08407284319400787, -0.048442400991916656, 0.11850221455097198, -0.12928882241249084, 0.03510063886642456, 0.09072701632976532, -0.3059840500354767, 0.03254007548093796, -0.008741389960050583, 0.1082751601934433, 0.030528780072927475, -0.0180516354739666, 0.04416965693235397, 0.039800647646188736, 0.02906673587858677, -0.07198981195688248, -0.09686371684074402, -0.263102263212204, -0.04276582598686218, -0.021692734211683273, -0.03583285212516785, 0.2953556180000305, 0.013480016961693764, 0.0029703634791076183, -0.056763965636491776, 0.004751852247864008, 0.002488786354660988, -0.0234270840883255, -0.022950569167733192, -0.00011496758088469505, 0.09344629943370819, 0.06504546105861664, -0.014789998531341553, -0.12997548282146454, -0.03008251264691353, -0.16328859329223633, 0.1063593178987503, 0.010780636221170425, 0.05411107465624809, -0.11326847970485687, -0.021179324015975, -0.022943520918488503, -0.13907673954963684, -0.029781993478536606, 0.02993990108370781, 0.05650787428021431, 0.11204512417316437, -0.027200469747185707, 0.0038838826585561037, 0.17308425903320312, 0.0587666817009449, 0.04662638157606125, 0.03477095440030098, -0.07078542560338974, 0.11083803325891495, -0.0016037926543504, 0.08872164785861969, -0.09717798233032227, -0.016405750066041946, 0.03292255848646164, -0.07716850936412811, 0.10516884177923203, -0.06375996023416519, -0.09949694573879242, -0.0414380207657814, 0.045336104929447174, 0.09893336147069931, 0.026906972751021385, 0.01120441872626543, -0.03387709707021713, 0.017556071281433105, 0.08342943340539932, -0.09677831083536148, -0.018126750364899635, 0.005340370815247297, 0.04175899177789688, 0.1502186357975006, 0.02869986556470394, 0.06971611082553864, -0.05754847079515457, 0.017681816592812538, 0.03533018007874489, 0.01365895289927721, 0.04295112192630768, 0.011792817153036594, 0.07587540149688721, -0.09161505848169327, 0.04123806580901146, -0.13916847109794617, -0.05758368968963623, -0.02813274785876274, 0.005628899671137333, -0.011430628597736359, -0.08278893679380417, -0.02264423295855522, -0.03695983067154884, 0.058812547475099564, -0.1383151113986969, 0.04107329994440079, -0.09329671412706375, 0.08532483875751495, -0.0013651892077177763, 0.10123667865991592, -0.12948940694332123, 0.06992451101541519, -0.06052519753575325, 0.04198133572936058, 0.015822846442461014, 0.030214231461286545, -0.08007824420928955, 0.047936856746673584, -0.10055293142795563, -0.03678375855088234, -0.06328481435775757, 0.029428726062178612, -0.032187171280384064, 0.06563571095466614, -0.1818680763244629, -0.07770366966724396, 0.13850969076156616, -0.14375771582126617, -0.15263576805591583, 0.11747175455093384, 0.07669349759817123, 0.03524131327867508, -0.00020119323744438589, 0.3112424910068512, -0.05738155171275139, -0.09556926041841507, 0.004872411955147982, 0.09284032881259918, -0.031244784593582153, -0.10566370189189911, 0.0951586589217186, -0.06485994905233383, 0.024189990013837814, 0.045207250863313675, 0.049524981528520584, 0.07998334616422653, 0.003591985907405615, -0.08766381442546844, -0.01839139312505722, -0.10281389951705933, 0.0038906496483832598, -0.058030009269714355, 0.049988098442554474, -0.06445382535457611, 0.008163121528923512, -0.01741776056587696, 0.10639456659555435, -0.011847997084259987, 0.061757706105709076, -0.0984429121017456, 0.12421135604381561, -0.08167654275894165, 0.01371620874851942, -0.10816625505685806, 0.17857897281646729, -0.023691896349191666, 0.03169337660074234, 0.1035289540886879, 0.04176751524209976, 0.03207042068243027, -0.11722423881292343, 0.002835117280483246, -0.04760536551475525, 0.12901940941810608, 0.06895516067743301, 0.051076341420412064, -0.08890165388584137, 0.04375474900007248, -0.033414166420698166, -0.004495947156101465, 0.0469282791018486, -0.0639895424246788, 0.033362239599227905, 0.1030045747756958, -0.0494871512055397, 0.06476140022277832, 0.06367865949869156, 0.045997120440006256, -0.01727491430938244, 0.03242136538028717, 0.03386053442955017, 0.023687776178121567, -0.09213550388813019, 0.33129313588142395, -0.12485034763813019, 0.25760117173194885, 0.22602017223834991, -0.184463232755661, 0.0940319076180458, 0.15988317131996155, 0.0286673866212368, -0.04318586364388466, 0.022271135821938515, 0.0233779139816761, 0.09644553065299988, -0.03724730759859085, 0.10636276006698608, -0.07800984382629395, 0.0320458821952343, 0.031085997819900513, -0.11046173423528671, -0.056019674986600876, 0.02367558889091015, 0.004478156566619873, -0.0310315303504467, 0.15607042610645294, 0.20195439457893372, -0.08851902931928635, 0.12930937111377716, -0.06813617050647736, -0.009545370936393738, 0.014842103235423565, -0.013496503233909607, -0.05106692388653755, 0.06993065774440765, -0.31190693378448486, -0.05209255963563919, 0.10828755050897598, 0.05585610866546631, 0.07045680284500122, -0.13314250111579895, -0.053246110677719116, -0.026496823877096176, -0.049142271280288696, -0.08371859788894653, 0.08278369903564453, -0.023674746975302696, 0.07648750394582748, -0.09230099618434906, -0.11201725155115128, 0.03273051604628563, -0.0404399037361145, -0.06813528388738632, 0.06398848444223404, -0.12469319999217987, -0.2867194712162018, -0.104000523686409, -0.07283447682857513, -0.03624533861875534, 0.004629232920706272, 0.15239672362804413, -0.10632498562335968, -0.02321973815560341, 0.0072460719384253025, -0.09090130776166916, -0.04055462032556534, 0.023135822266340256, -0.01329275406897068, 0.013157078996300697, 0.02613654173910618, -0.1376633644104004, -0.011953582987189293, 0.010954131372272968, 0.010547241196036339, 0.08513173460960388, -0.007837070152163506, 0.07016739249229431, 0.1771816909313202, 0.07669303566217422, 0.010316437110304832, 0.011017982847988605, 0.10627076774835587, -0.11895402520895004, -0.056307900696992874, 0.20689861476421356, -0.032657600939273834, -0.008226466365158558, 0.20359304547309875, 0.02453932724893093, -0.060254912823438644, -0.016281381249427795, -0.057226233184337616, -0.06213725358247757, -0.250810444355011, -0.09473282843828201, -0.11870662868022919, -0.03488190099596977, -0.09614710509777069, 0.05402427166700363, 0.10478762537240982, -0.0028168957214802504, 0.01989085040986538, -0.032190121710300446, -0.0065746246837079525, -0.018913509324193, 0.20318889617919922, -0.054832346737384796, 0.10856698453426361, -0.10049797594547272, -0.05231170728802681, 0.05832316726446152, 0.14112453162670135, 0.0842883288860321, 0.17044255137443542, 0.06540867686271667, 0.09212805330753326, 0.20915718376636505, 0.1370765119791031, 0.03648846223950386, -0.010793140158057213, 0.018133599311113358, -0.021886643022298813, -0.05732676386833191, -0.02956974506378174, 0.037365078926086426, 0.1598787158727646, -0.0748477578163147, 0.02857080101966858, -0.15312500298023224, 0.05543497949838638, 0.1433180868625641, 0.07259305566549301, -0.05521240830421448, 0.03402937576174736, 0.06275461614131927, -0.056927215307950974, -0.0002897959202528, 0.09465410560369492, 0.0758315920829773, -0.09155488014221191, 0.10950776189565659, 0.11049742996692657, 0.06760648638010025, -0.08094844967126846, 0.03977765142917633, -0.13752439618110657, -0.15809990465641022, 0.07091164588928223, 0.0885554701089859, -0.2648575007915497, 0.26643654704093933, 0.0007909201085567474, -0.002856183797121048, -0.0647963434457779, 0.01712317205965519, 0.038931675255298615, 0.06283941119909286, 0.1181614026427269, 0.03156203776597977, -0.12337813526391983, -0.027605298906564713, -0.0616597905755043, 0.06341032683849335, 0.058567874133586884, 0.1429598480463028, -0.1038546934723854, 0.012197349220514297, -0.02168693207204342, 0.0021959892474114895, 0.049749720841646194, -0.12424419075250626, -0.09012389928102493, 0.04187798872590065, 0.25944915413856506, 0.08418704569339752, -0.04109339788556099, -0.05393208935856819, -0.13151904940605164, 0.0814642682671547, -0.07047543674707413, -0.015531527809798717, -0.03914453461766243, -0.22496695816516876, 0.1434544026851654, -0.04508580267429352, 0.020882677286863327, -0.027470558881759644, -0.03225952014327049, -0.05052776262164116, -0.06405045092105865, 0.133650004863739, -0.08850934356451035, -0.05544538050889969, -0.03956685960292816, 0.1888236254453659, -0.05095597356557846, 0.06897120922803879, 0.04447295516729355, 0.0045534223318099976, -0.045498453080654144, -0.08772400766611099, 0.04806334525346756, 0.0779615193605423, -0.06662064045667648, 0.019613629207015038, 0.039448220282793045, -0.1939699649810791, -0.02691502869129181, -0.11337212473154068, 0.1984068900346756, 0.08627109229564667, -0.020082730799913406, 0.17588847875595093, 0.2831975817680359, -0.039837516844272614, -0.24749970436096191, -0.22726121544837952, -0.07637999206781387, -0.020427441224455833, -0.09177924692630768, -0.10419391095638275, 0.08422007411718369, 0.0229727141559124, -0.11597783863544464, -0.017542202025651932, -0.19328416883945465, -0.11392750591039658, 0.1937478929758072, -0.09531806409358978, 0.27122461795806885, -0.11962014436721802, -0.12441156059503555, -0.05875333026051521, -0.16312174499034882, 0.042483218014240265, -0.08885679394006729, 0.07732022553682327, 0.034149643033742905, 0.05970950797200203, -0.003287240397185087, -0.003827819135040045, 0.20141272246837616, 0.04430588707327843, -0.042263489216566086, -0.061027318239212036, 0.002349474932998419, 0.022575901821255684, 0.03008701279759407, 0.08596446365118027, -0.1389489769935608, -0.005027602426707745, -0.10646910965442657, -0.020905543118715286, -0.117137610912323, 0.08184660971164703, 0.06619961559772491, -0.017289891839027405, -0.004916450474411249, -0.05005017668008804, -0.0034686692524701357, 0.03442664071917534, 0.2297755777835846, -0.02876785770058632, 0.03053349256515503, 0.1517728865146637, 0.10158368200063705, -0.1957353800535202, 0.02506314218044281, -0.10115732997655869, -0.07449017465114594, 0.09610037505626678, -0.10843272507190704, 0.02265559509396553, 0.030077164992690086, -0.02379048801958561, 0.08058442175388336, 0.02216428332030773, -0.10025763511657715, 0.11071550846099854, 0.10957606136798859, -0.0176381915807724, -0.07819406688213348, 0.04133215546607971, 0.08696489781141281, 0.1291920691728592, 0.03256852179765701, 0.09220968186855316, -0.039396174252033234, -0.027809834107756615, -0.014642981812357903, 0.01947222277522087, -0.17124584317207336, 0.133036270737648, 0.039687830954790115, 0.05455479770898819, -0.14383593201637268, 0.1618543118238449, 0.05160878226161003, -0.13928291201591492, 0.038216136395931244, 0.025930076837539673, -0.06438613682985306, -0.11165821552276611, -0.04855933412909508, 0.019988808780908585, 0.016058314591646194, -0.17280754446983337, 0.013601422309875488, -0.14431551098823547, -0.005042436998337507, 0.08755294978618622, 0.04361502081155777, 0.05723707750439644, -0.02894662879407406, -0.06780143082141876, 0.07119295746088028, -0.029338568449020386, -0.002023112028837204, 0.017209593206644058, -0.12056823074817657, -0.07790213823318481, 0.11192914098501205, 0.12586674094200134, -0.06054440885782242, -0.08830146491527557, -0.07112592458724976, 0.06316087394952774, -0.12042523920536041, 0.020623886957764626, -0.13937032222747803, -0.017077788710594177, 0.020429451018571854, -0.1441761702299118, -0.032389093190431595, 0.010545270517468452, -0.10369016975164413, 0.0003367133322171867, -0.008959678933024406, 0.09425868839025497, -0.12766221165657043, -0.029141897335648537, 0.07413143664598465, -0.014195372350513935, 0.10658973455429077, 0.16998344659805298, -0.10709546506404877, 0.11193473637104034, -0.16355225443840027, -0.17495086789131165, 0.1551673710346222, 0.06059209629893303, 0.00829382985830307, -0.055971186608076096, -0.02159471996128559, 0.1589343249797821, 0.056846387684345245, 0.0096755875274539, 0.07413005083799362, -0.1018596887588501, -0.07454317808151245, -0.12277984619140625, -0.07211081683635712, -0.008348295465111732, -0.053984951227903366, 0.216748908162117, 0.03366713970899582, 0.09887143224477768, -0.04048221558332443, 0.016102056950330734, -0.04049437493085861, 0.0407869815826416, -0.10530024766921997, -0.1500396728515625, -0.09838276356458664, -0.04455193132162094, -0.013451064005494118, -0.0634477511048317, 0.23963245749473572, -0.025219416245818138, -0.04346635192632675, 0.0029004602693021297, 0.02877218648791313, -0.10544140636920929, 0.03396975249052048, 0.3093136250972748, 0.10150681436061859, -0.004045445006340742, -0.053342919796705246, -0.06757697463035583, 0.007496279664337635, 0.04705372452735901, -0.09847608953714371, 0.167817622423172, 0.16348469257354736, 0.12396731227636337, 0.144182488322258, -0.028735369443893433, -0.13156335055828094, -0.10512306541204453, -0.04924098402261734, 0.05977170541882515, -0.018139639869332314, 0.1184292584657669, 0.14625157415866852, -0.018524032086133957, 0.03380046784877777, -0.03137396648526192, -0.008397310972213745, -0.1425568163394928, -0.15760569274425507, -0.07288917899131775, -0.12968841195106506, -0.009307464584708214, -0.03224848955869675, 0.056581296026706696, 0.1044769287109375, 0.05450006201863289, -0.034661851823329926, 0.03454272821545601, -0.03445551171898842, -0.045392878353595734, 0.06270688772201538, -0.036125410348176956, -0.012343024834990501, -0.06476598232984543, -0.01762300543487072, 0.030668627470731735, 0.04740387946367264, -0.01688678190112114, 0.03550167754292488, -0.09875012934207916, 0.024105703458189964, -0.12057030200958252, -0.08665617555379868, -0.03458499163389206, -0.03779974579811096, 0.0413118340075016, 0.15594364702701569, 0.0746096521615982, -0.021158510819077492, 0.069725900888443, 0.18407267332077026, -0.08445937931537628, -0.17723152041435242, -0.06324465572834015, 0.06939052045345306, -0.042912282049655914, 0.037337616086006165, -0.028433848172426224, -0.03128033131361008, -0.10462261736392975, 0.26729637384414673, 0.23264947533607483, -0.018240921199321747, 0.07124070823192596, -0.021166015416383743, 0.025142615661025047, -0.06741426885128021, -0.0011614372488111258, 0.1813926249742508, 0.1594226062297821, -0.014077847823500633, 0.000804136332590133, -0.10039128363132477, -0.05501105263829231, -0.09051047265529633, 0.05251096561551094, -0.05440937355160713, -0.14800594747066498, -0.002132957801222801, 0.10825470834970474, -0.20405754446983337, 0.026306480169296265, -0.08346038311719894, -0.1528240144252777, -0.09594543278217316, 0.005131426267325878, 0.14664645493030548, 0.17241241037845612, -0.033344466239213943, -0.031029364094138145, -0.017430324107408524, 0.08082691580057144, 0.0317678265273571, -0.15730194747447968, -0.03345586732029915, 0.02667984366416931, -0.04289327934384346, 0.049868740141391754, 0.036087073385715485, 0.07046983391046524, 0.01797877438366413, 0.16522200405597687, -0.044321052730083466, 0.20048654079437256, 0.052845634520053864, -0.048490170389413834, -0.004364688415080309, 0.08587443083524704, -0.0029430598951876163, 0.023948458954691887, 0.06607846915721893, -0.08147697150707245, 0.05333482474088669, 0.007816185243427753, 0.0037135384045541286, -0.07621578872203827, -0.0012713692849501967, -0.06116130203008652, 0.05253646522760391, -0.041140370070934296, -0.047523654997348785, -0.05986494570970535, -0.010361285880208015, 0.020688015967607498, -0.005675795488059521, -0.14720213413238525, -0.10264343023300171, -0.09013962745666504, -0.04317831993103027, -0.08774956315755844, 0.06394899636507034, -0.06829378753900528, 0.012464096769690514, -0.0634537860751152, 0.0038461522199213505, -0.03730455040931702, -0.017608152702450752, 0.059453338384628296, -0.003483641194179654, 0.0064596445299685, -0.08771581947803497, 0.10352864116430283, 0.07544658333063126, -0.11591224372386932, -0.0659467801451683 ]
null
null
transformers
~~~ # requirement packages !pip install git+https://github.com/huggingface/datasets.git !pip install git+https://github.com/huggingface/transformers.git !pip install torchaudio !pip install librosa !git clone https://github.com/m3hrdadfi/soxan cd soxan ~~~ # prediction ~~~ import torch import torch.nn as nn import torch.nn.functional as F import torchaudio from transformers import AutoConfig, Wav2Vec2FeatureExtractor import librosa import IPython.display as ipd import numpy as np import pandas as pd ~~~ ~~~ device = torch.device("cuda" if torch.cuda.is_available() else "cpu") model_name_or_path = "Bagus/wav2vec2-xlsr-greek-speech-emotion-recognition" config = AutoConfig.from_pretrained(model_name_or_path) feature_extractor = Wav2Vec2FeatureExtractor.from_pretrained(model_name_or_path) sampling_rate = feature_extractor.sampling_rate model = Wav2Vec2ForSpeechClassification.from_pretrained(model_name_or_path).to(device) ~~~ ~~~ def speech_file_to_array_fn(path, sampling_rate): speech_array, _sampling_rate = torchaudio.load(path) resampler = torchaudio.transforms.Resample(_sampling_rate) speech = resampler(speech_array).squeeze().numpy() return speech def predict(path, sampling_rate): speech = speech_file_to_array_fn(path, sampling_rate) inputs = feature_extractor(speech, sampling_rate=sampling_rate, return_tensors="pt", padding=True) inputs = {key: inputs[key].to(device) for key in inputs} with torch.no_grad(): logits = model(**inputs).logits scores = F.softmax(logits, dim=1).detach().cpu().numpy()[0] outputs = [{"Emotion": config.id2label[i], "Score": f"{round(score * 100, 3):.1f}%"} for i, score in enumerate(scores)] return outputs ~~~ # prediction ~~~ # path for a sample path = '/data/jtes_v1.1/wav/f01/ang/f01_ang_01.wav' outputs = predict(path, sampling_rate) ~~~ ~~~ [{'Emotion': 'anger', 'Score': '98.3%'}, {'Emotion': 'disgust', 'Score': '0.0%'}, {'Emotion': 'fear', 'Score': '0.4%'}, {'Emotion': 'happiness', 'Score': '0.7%'}, {'Emotion': 'sadness', 'Score': '0.5%'}] ~~~
{"language": "el", "license": "apache-2.0", "tags": ["audio", "audio-classification", "speech"], "datasets": ["aesdd"]}
audio-classification
Bagus/wav2vec2-xlsr-greek-speech-emotion-recognition
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "audio", "audio-classification", "speech", "el", "dataset:aesdd", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[ "el" ]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #audio #audio-classification #speech #el #dataset-aesdd #license-apache-2.0 #endpoints_compatible #region-us
~~~ # requirement packages !pip install git+URL !pip install git+URL !pip install torchaudio !pip install librosa !git clone URL cd soxan ~~~ # prediction ~~~ import torch import URL as nn import URL.functional as F import torchaudio from transformers import AutoConfig, Wav2Vec2FeatureExtractor import librosa import IPython.display as ipd import numpy as np import pandas as pd ~~~ ~~~ device = URL("cuda" if URL.is_available() else "cpu") model_name_or_path = "Bagus/wav2vec2-xlsr-greek-speech-emotion-recognition" config = AutoConfig.from_pretrained(model_name_or_path) feature_extractor = Wav2Vec2FeatureExtractor.from_pretrained(model_name_or_path) sampling_rate = feature_extractor.sampling_rate model = Wav2Vec2ForSpeechClassification.from_pretrained(model_name_or_path).to(device) ~~~ ~~~ def speech_file_to_array_fn(path, sampling_rate): speech_array, _sampling_rate = URL(path) resampler = torchaudio.transforms.Resample(_sampling_rate) speech = resampler(speech_array).squeeze().numpy() return speech def predict(path, sampling_rate): speech = speech_file_to_array_fn(path, sampling_rate) inputs = feature_extractor(speech, sampling_rate=sampling_rate, return_tensors="pt", padding=True) inputs = {key: inputs[key].to(device) for key in inputs} with torch.no_grad(): logits = model(inputs).logits scores = F.softmax(logits, dim=1).detach().cpu().numpy()[0] outputs = [{"Emotion": config.id2label[i], "Score": f"{round(score * 100, 3):.1f}%"} for i, score in enumerate(scores)] return outputs ~~~ # prediction ~~~ # path for a sample path = '/data/jtes_v1.1/wav/f01/ang/f01_ang_01.wav' outputs = predict(path, sampling_rate) ~~~ ~~~ [{'Emotion': 'anger', 'Score': '98.3%'}, {'Emotion': 'disgust', 'Score': '0.0%'}, {'Emotion': 'fear', 'Score': '0.4%'}, {'Emotion': 'happiness', 'Score': '0.7%'}, {'Emotion': 'sadness', 'Score': '0.5%'}] ~~~
[ "# requirement packages\n!pip install git+URL\n!pip install git+URL\n!pip install torchaudio\n!pip install librosa\n!git clone URL\ncd soxan\n~~~", "# prediction\n~~~\nimport torch\nimport URL as nn\nimport URL.functional as F\nimport torchaudio\nfrom transformers import AutoConfig, Wav2Vec2FeatureExtractor\n\nimport librosa\nimport IPython.display as ipd\nimport numpy as np\nimport pandas as pd\n~~~\n\n~~~\ndevice = URL(\"cuda\" if URL.is_available() else \"cpu\")\nmodel_name_or_path = \"Bagus/wav2vec2-xlsr-greek-speech-emotion-recognition\"\nconfig = AutoConfig.from_pretrained(model_name_or_path)\nfeature_extractor = Wav2Vec2FeatureExtractor.from_pretrained(model_name_or_path)\nsampling_rate = feature_extractor.sampling_rate\nmodel = Wav2Vec2ForSpeechClassification.from_pretrained(model_name_or_path).to(device)\n~~~\n\n~~~\ndef speech_file_to_array_fn(path, sampling_rate):\n speech_array, _sampling_rate = URL(path)\n resampler = torchaudio.transforms.Resample(_sampling_rate)\n speech = resampler(speech_array).squeeze().numpy()\n return speech\n\n\ndef predict(path, sampling_rate):\n speech = speech_file_to_array_fn(path, sampling_rate)\n inputs = feature_extractor(speech, sampling_rate=sampling_rate, return_tensors=\"pt\", padding=True)\n inputs = {key: inputs[key].to(device) for key in inputs}\n\n with torch.no_grad():\n logits = model(inputs).logits\n\n scores = F.softmax(logits, dim=1).detach().cpu().numpy()[0]\n outputs = [{\"Emotion\": config.id2label[i], \"Score\": f\"{round(score * 100, 3):.1f}%\"} for i, score in enumerate(scores)]\n return outputs\n~~~", "# prediction\n~~~", "# path for a sample\npath = '/data/jtes_v1.1/wav/f01/ang/f01_ang_01.wav' \noutputs = predict(path, sampling_rate)\n~~~\n\n~~~\n[{'Emotion': 'anger', 'Score': '98.3%'},\n {'Emotion': 'disgust', 'Score': '0.0%'},\n {'Emotion': 'fear', 'Score': '0.4%'},\n {'Emotion': 'happiness', 'Score': '0.7%'},\n {'Emotion': 'sadness', 'Score': '0.5%'}]\n ~~~" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #audio #audio-classification #speech #el #dataset-aesdd #license-apache-2.0 #endpoints_compatible #region-us \n", "# requirement packages\n!pip install git+URL\n!pip install git+URL\n!pip install torchaudio\n!pip install librosa\n!git clone URL\ncd soxan\n~~~", "# prediction\n~~~\nimport torch\nimport URL as nn\nimport URL.functional as F\nimport torchaudio\nfrom transformers import AutoConfig, Wav2Vec2FeatureExtractor\n\nimport librosa\nimport IPython.display as ipd\nimport numpy as np\nimport pandas as pd\n~~~\n\n~~~\ndevice = URL(\"cuda\" if URL.is_available() else \"cpu\")\nmodel_name_or_path = \"Bagus/wav2vec2-xlsr-greek-speech-emotion-recognition\"\nconfig = AutoConfig.from_pretrained(model_name_or_path)\nfeature_extractor = Wav2Vec2FeatureExtractor.from_pretrained(model_name_or_path)\nsampling_rate = feature_extractor.sampling_rate\nmodel = Wav2Vec2ForSpeechClassification.from_pretrained(model_name_or_path).to(device)\n~~~\n\n~~~\ndef speech_file_to_array_fn(path, sampling_rate):\n speech_array, _sampling_rate = URL(path)\n resampler = torchaudio.transforms.Resample(_sampling_rate)\n speech = resampler(speech_array).squeeze().numpy()\n return speech\n\n\ndef predict(path, sampling_rate):\n speech = speech_file_to_array_fn(path, sampling_rate)\n inputs = feature_extractor(speech, sampling_rate=sampling_rate, return_tensors=\"pt\", padding=True)\n inputs = {key: inputs[key].to(device) for key in inputs}\n\n with torch.no_grad():\n logits = model(inputs).logits\n\n scores = F.softmax(logits, dim=1).detach().cpu().numpy()[0]\n outputs = [{\"Emotion\": config.id2label[i], \"Score\": f\"{round(score * 100, 3):.1f}%\"} for i, score in enumerate(scores)]\n return outputs\n~~~", "# prediction\n~~~", "# path for a sample\npath = '/data/jtes_v1.1/wav/f01/ang/f01_ang_01.wav' \noutputs = predict(path, sampling_rate)\n~~~\n\n~~~\n[{'Emotion': 'anger', 'Score': '98.3%'},\n {'Emotion': 'disgust', 'Score': '0.0%'},\n {'Emotion': 'fear', 'Score': '0.4%'},\n {'Emotion': 'happiness', 'Score': '0.7%'},\n {'Emotion': 'sadness', 'Score': '0.5%'}]\n ~~~" ]
[ 60, 42, 522, 5, 165 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #audio #audio-classification #speech #el #dataset-aesdd #license-apache-2.0 #endpoints_compatible #region-us \n# requirement packages\n!pip install git+URL\n!pip install git+URL\n!pip install torchaudio\n!pip install librosa\n!git clone URL\ncd soxan\n~~~" ]
[ -0.08011195063591003, 0.14450164139270782, -0.005670858547091484, 0.040525052696466446, 0.14206655323505402, -0.0022708692122250795, 0.06237025931477547, 0.08335468173027039, -0.0055440496653318405, -0.008632436394691467, 0.061570484191179276, 0.20992588996887207, 0.07769520580768585, -0.00413459911942482, -0.008608619682490826, -0.24474215507507324, 0.008546733297407627, 0.04790317639708519, 0.001423539943061769, 0.09386266767978668, 0.07928725332021713, 0.005584997590631247, 0.05131334811449051, 0.021929314360022545, -0.07599183917045593, 0.014699341729283333, 0.03189966455101967, -0.11150036007165909, 0.022309724241495132, 0.034482959657907486, 0.08381403982639313, 0.023974386975169182, 0.007020864635705948, -0.1280113309621811, 0.026246797293424606, 0.05326694995164871, 0.015727974474430084, 0.07067689299583435, 0.017523981630802155, -0.004591343924403191, -0.007144948933273554, 0.01480917539447546, -0.06188800185918808, 0.027453910559415817, -0.05110229179263115, -0.141433984041214, -0.0740094780921936, 0.09851976484060287, 0.04642505198717117, 0.10036101192235947, 0.019225601106882095, 0.1687057614326477, -0.007820358499884605, 0.05679764971137047, 0.2902436852455139, -0.21123242378234863, -0.028728768229484558, -0.02364981174468994, 0.16956904530525208, 0.1389685869216919, 0.020301125943660736, 0.033080678433179855, 0.06396075338125229, 0.03769950568675995, -0.057075221091508865, -0.08630497753620148, -0.16116738319396973, -0.028465626761317253, -0.1161428838968277, 0.0013022056082263589, 0.2168128937482834, -0.031861480325460434, -0.061865270137786865, 0.06720277667045593, -0.03506358712911606, -0.10307363420724869, -0.026197858154773712, 0.06973870098590851, -0.005461561027914286, 0.02585766464471817, -0.014886069111526012, -0.1294206827878952, -0.06669566035270691, -0.08136521279811859, -0.03975660353899002, 0.14649876952171326, 0.057666242122650146, 0.05525749549269676, -0.03787161409854889, 0.062212973833084106, -0.03406576067209244, -0.10715501010417938, -0.033401377499103546, -0.04872380569577217, 0.053193096071481705, 0.023076916113495827, 0.0020844608079642057, -0.02536366507411003, 0.0756467953324318, 0.07333597540855408, -0.013381971046328545, 0.024248598143458366, -0.01659313030540943, 0.10919703543186188, -0.019424643367528915, 0.01962076872587204, -0.11200445890426636, -0.021822530776262283, 0.09294965118169785, -0.053066521883010864, 0.1254296749830246, -0.015317286364734173, -0.11592864990234375, 0.01865755021572113, -0.08035793155431747, 0.07154198735952377, 0.0681774765253067, -0.018496183678507805, -0.06538324803113937, -0.007728792726993561, 0.09189290553331375, -0.034747857600450516, 0.00024602352641522884, -0.0388001874089241, 0.025727566331624985, 0.05845336616039276, 0.08326243609189987, 0.047567687928676605, -0.07391656935214996, 0.09562816470861435, 0.012261766009032726, -0.010399271734058857, -0.011762273497879505, -0.0438246876001358, 0.09818405658006668, -0.1429646909236908, 0.08094917237758636, -0.0898064374923706, -0.06956025213003159, 0.0648207738995552, 0.07658836990594864, 0.03206966444849968, -0.14068306982517242, 0.10930173844099045, -0.0560372956097126, -0.016468247398734093, -0.08383949100971222, 0.017133286222815514, -0.06653548777103424, 0.10327570140361786, 0.008348597213625908, 0.0605241023004055, -0.12736818194389343, 0.0489388033747673, -0.05019417777657509, 0.03187311440706253, -0.24998384714126587, -0.02009543403983116, -0.07647265493869781, 0.006424384657293558, -0.07846707850694656, -0.11031164228916168, -0.03900827467441559, 0.05617482215166092, 0.043942891061306, 0.03733310475945473, -0.08195522427558899, -0.06882388889789581, 0.16355440020561218, -0.11868657171726227, -0.08770701289176941, 0.12205123901367188, 0.07961204648017883, -0.001707881921902299, 0.0624614879488945, 0.1730031669139862, 0.1526826173067093, -0.25423479080200195, -0.009004333056509495, 0.11055225878953934, -0.051084477454423904, -0.16109338402748108, 0.10441970825195312, -0.02198103256523609, -0.027934938669204712, 0.009991563856601715, -0.07450126856565475, 0.08559738099575043, -0.026190565899014473, -0.041929297149181366, -0.05372948944568634, -0.0926019623875618, -0.10513736307621002, -0.0063578723929822445, -0.05155407264828682, -0.0006014693644829094, -0.06712371110916138, 0.008110814727842808, 0.07122418284416199, -0.00961250439286232, 0.05466604232788086, -0.024754967540502548, 0.10409341007471085, -0.05713654309511185, -0.012068411335349083, -0.04854718595743179, 0.11839249730110168, -0.008547269739210606, 0.011385289952158928, 0.13463295996189117, -0.016573801636695862, 0.019516780972480774, -0.028448082506656647, -0.03099401295185089, -0.024047471582889557, 0.07159125059843063, -0.008725172840058804, -0.019342966377735138, -0.1261204183101654, 0.030539460480213165, 0.006472480017691851, 0.0227329283952713, 0.01147555187344551, -0.009124381467700005, 0.017574507743120193, 0.044122084975242615, -0.030137570574879646, -0.008971183560788631, 0.051169395446777344, -0.010190906003117561, -0.02977670356631279, -0.013074557296931744, 0.06809116899967194, 0.039682429283857346, -0.05399096757173538, 0.12532860040664673, 0.006183347199112177, 0.06954792141914368, 0.18238818645477295, -0.12487997859716415, 0.05516795068979263, 0.07319317013025284, -0.014867703430354595, -0.036986321210861206, -0.006478822790086269, -0.06207960471510887, 0.0796104446053505, 0.01615612767636776, 0.08591161668300629, -0.07109588384628296, 0.023765485733747482, 0.06616511940956116, -0.07577759027481079, 0.03889484703540802, 0.050890181213617325, 0.12690110504627228, -0.020598476752638817, 0.11403746157884598, 0.13043709099292755, -0.13392968475818634, 0.11969240754842758, -0.028723401948809624, -0.03859836608171463, 0.013636691495776176, 0.04629818722605705, 0.029567137360572815, 0.11517064273357391, -0.12157601118087769, 0.042313020676374435, 0.05842069908976555, -0.036959730088710785, 0.009197843261063099, -0.14930526912212372, -0.05748153105378151, 0.008592505007982254, -0.06362588703632355, -0.14812619984149933, -0.010503936558961868, -0.04469863697886467, 0.022786490619182587, -0.0819615051150322, -0.05198507383465767, 0.054845403879880905, 0.028511596843600273, -0.05870714411139488, 0.07028599083423615, -0.12698519229888916, -0.194029301404953, -0.19337770342826843, -0.14077694714069366, 0.0046616205945611, 0.05091341584920883, 0.13599053025245667, -0.06201408430933952, 0.0004600401734933257, 0.04531640559434891, 0.08083542436361313, -0.07666337490081787, 0.005762027110904455, -0.05947665497660637, 0.0645429864525795, 0.04116826131939888, -0.08575639128684998, 0.018979983404278755, 0.016306662932038307, 0.002185208722949028, 0.02314578741788864, 0.018737101927399635, 0.06222664192318916, 0.07743817567825317, 0.0800318568944931, -0.017406970262527466, -0.02940337173640728, 0.15392650663852692, -0.10218168795108795, 0.0702502503991127, 0.21654346585273743, -0.087334543466568, 0.03913174942135811, 0.15929068624973297, 0.051169343292713165, 0.027264606207609177, -0.0201253779232502, 0.00714898994192481, -0.11734043806791306, -0.32702600955963135, -0.020797910168766975, -0.09075858443975449, 0.06637565046548843, -0.0107341343536973, 0.03848671168088913, -0.006008605472743511, 0.1106480285525322, -0.08357907831668854, 0.02304789237678051, 0.01374572142958641, 0.0007996202330105007, 0.06693466007709503, -0.07496945559978485, 0.07891668379306793, -0.04381921514868736, -0.04630392789840698, 0.09615079313516617, 0.1867498755455017, 0.14772014319896698, 0.08127884566783905, 0.14688946306705475, 0.07193545997142792, 0.07102952152490616, 0.083553746342659, 0.08000489324331284, 0.02308303490281105, 0.02417684718966484, -0.0077904462814331055, -0.08315446972846985, 0.04014473035931587, 0.0038290945813059807, 0.01864505372941494, -0.07152841985225677, 0.0319003164768219, -0.04119284078478813, 0.00639046635478735, 0.22280816733837128, -0.029437769204378128, -0.11930934339761734, 0.03837106376886368, 0.041890766471624374, 0.031148679554462433, -0.05238587036728859, 0.05269623547792435, 0.0067258188501000404, -0.0111433370038867, 0.10288403183221817, -0.013403386808931828, 0.11756668239831924, -0.06968046724796295, -0.022930579259991646, 0.08427078276872635, 0.10964860767126083, 0.03949020802974701, 0.06048760190606117, -0.13697189092636108, 0.14035965502262115, 0.0559556670486927, 0.06163892522454262, -0.010061192326247692, 0.030454004183411598, -0.004538698587566614, 0.09617583453655243, 0.1782284379005432, 0.03337865322828293, -0.04464201629161835, -0.07212352007627487, -0.09661649912595749, 0.018038857728242874, 0.025398435071110725, 0.06777647137641907, -0.06364157050848007, 0.0062468028627336025, -0.02378562092781067, 0.011811130680143833, -0.10813289880752563, -0.1354084610939026, -0.10463783890008926, 0.03523140400648117, 0.17606741189956665, -0.03762542083859444, -0.06266363710165024, -0.044165272265672684, -0.1594492346048355, 0.19963429868221283, -0.14094656705856323, -0.10704182088375092, -0.017723161727190018, -0.04835943877696991, 0.09819018095731735, -0.023743106052279472, 0.031003808602690697, -0.06835345923900604, -0.020951950922608376, -0.06421675533056259, -0.0637330636382103, 0.07645151764154434, -0.11028552800416946, -0.027043908834457397, -0.04182123765349388, 0.249094158411026, -0.029959579929709435, 0.051098864525556564, -0.05921478569507599, -0.02280447632074356, -0.02403601072728634, -0.0981733649969101, -0.03713783621788025, -0.025400662794709206, -0.026516050100326538, 0.0726826936006546, -0.1212533712387085, 0.002112187212333083, -0.020593231543898582, -0.08947626501321793, 0.15441429615020752, 0.10400751978158951, -0.0524393655359745, 0.04796221852302551, 0.13879534602165222, -0.09649429470300674, -0.24646393954753876, -0.06995252519845963, 0.02482985518872738, -0.07419935613870621, 0.024890223518013954, -0.2013607770204544, 0.1437625139951706, 0.09118098020553589, -0.033184900879859924, 0.1796603798866272, -0.22212304174900055, -0.03503446653485298, 0.06506594270467758, -0.01398635283112526, 0.14556826651096344, -0.14401906728744507, -0.09032736718654633, -0.07084894925355911, -0.13243263959884644, 0.14170017838478088, -0.20134547352790833, 0.11256474256515503, 0.05350630730390549, 0.056760434061288834, -0.0012747907312586904, -0.03984161093831062, 0.08127790689468384, -0.014985649846494198, -0.08826107531785965, -0.0050113690085709095, 0.07654311507940292, 0.02331939898431301, 0.0431111641228199, 0.04101070389151573, -0.16539277136325836, -0.029528804123401642, -0.015749214217066765, -0.022111250087618828, -0.04735357686877251, 0.11940488964319229, -0.007675147615373135, -0.008684719912707806, -0.07843159139156342, -0.08788643032312393, -0.039605870842933655, -0.001957772532477975, 0.18784652650356293, 0.008995757438242435, -0.0322880893945694, 0.11051098257303238, -0.011686529964208603, -0.02487538941204548, -0.15549147129058838, -0.011176184751093388, -0.0684714987874031, 0.08709675073623657, -0.06656670570373535, 0.016694333404302597, 0.06825447827577591, 0.012955890968441963, -0.020197290927171707, 0.031629737466573715, -0.08731997013092041, -0.019630001857876778, 0.04295165091753006, -0.1237151101231575, 0.006401074584573507, -0.03476100042462349, 0.09679949283599854, 0.021860096603631973, 0.06109233945608139, 0.14094890654087067, -0.04110132157802582, -0.028781525790691376, 0.04577105864882469, 0.008175292983651161, -0.17579272389411926, 0.1732199490070343, 0.08164554834365845, 0.011189315468072891, -0.1395411491394043, 0.09551619738340378, 0.007675327826291323, -0.13771142065525055, -0.03877026215195656, 0.056974880397319794, -0.04422665387392044, -0.11768504232168198, 0.027784675359725952, -0.10434693843126297, -0.0434427335858345, -0.12356032431125641, -0.012103807181119919, -0.0621492862701416, -0.008722521364688873, 0.020802035927772522, 0.043312620371580124, -0.033205099403858185, 0.017679814249277115, -0.0011446671560406685, -0.03234342858195305, 0.040193844586610794, -0.045781221240758896, 0.10978120565414429, -0.20447984337806702, -0.07786364108324051, 0.056093476712703705, 0.07789381593465805, -0.032516393810510635, 0.05240519717335701, -0.10482709854841232, 0.07561686635017395, 0.0042795948684215546, 0.053147125989198685, -0.046512771397829056, -0.029137544333934784, -0.005600589793175459, -0.047681115567684174, -0.019116055220365524, 0.05496485158801079, -0.03925628960132599, -0.01922147534787655, -0.001607144600711763, 0.07292013615369797, -0.093069888651371, -0.020494326949119568, 0.045745041221380234, -0.037541650235652924, 0.05936851724982262, 0.08874187618494034, -0.0738581046462059, 0.0019090468995273113, -0.15363705158233643, -0.03710725158452988, 0.12597781419754028, 0.0334087572991848, -0.018573923036456108, -0.054437775164842606, 0.06561440974473953, 0.05144403874874115, 0.03240962699055672, -0.028379525989294052, 0.15394560992717743, -0.1126813292503357, -0.07371281087398529, -0.01769424043595791, -0.056190475821495056, 0.011658916249871254, -0.046359576284885406, 0.21100783348083496, 0.048263803124427795, 0.12420040369033813, 0.01478149089962244, 0.012111942283809185, -0.08319493383169174, 0.0729527547955513, -0.02078997902572155, -0.12564149498939514, -0.03483758866786957, -0.005418259184807539, -0.01190694235265255, -0.006159237120300531, 0.1521865576505661, -0.025208627805113792, 0.01768467202782631, 0.01766630820930004, 0.09420705586671829, 0.005309601780027151, 0.015204653143882751, 0.17757754027843475, 0.008020365610718727, 0.00665949797257781, -0.11903707683086395, -0.06988777965307236, 0.067713662981987, -0.08638190478086472, -0.06720968335866928, 0.1115991473197937, 0.10912486165761948, 0.03794562444090843, 0.05123240873217583, -0.015017611905932426, -0.16824693977832794, -0.09569389373064041, 0.06550579518079758, 0.07951082289218903, 0.05163171887397766, 0.20106986165046692, 0.06686490774154663, -0.01711951196193695, 0.02733290009200573, -0.0005107876495458186, -0.06322646141052246, -0.1450149267911911, -0.08798503130674362, -0.045259028673172, -0.11298491060733795, 0.017013082280755043, -0.0284736230969429, 0.030974233523011208, 0.0029906847048550844, 0.011531255207955837, -0.05354203283786774, 0.11137641221284866, 0.02909880131483078, -0.10448620468378067, 0.012162826023995876, 0.03579520434141159, -0.03165693208575249, 0.030017593875527382, -0.03780198097229004, 0.06743375211954117, -0.006086669396609068, 0.007555673830211163, 0.011632529087364674, 0.0049606855027377605, 0.08581824600696564, -0.0984535962343216, -0.0404961034655571, -0.032895468175411224, 0.042099256068468094, 0.06374260783195496, 0.20461523532867432, -0.014127757400274277, -0.0021908406633883715, 0.044509805738925934, 0.08811858296394348, -0.0559069849550724, -0.04095767065882683, -0.06222027167677879, 0.11648882925510406, -0.09525223821401596, -0.0004955808399245143, -0.038029707968235016, -0.025530166923999786, -0.07802815735340118, 0.3282807469367981, 0.2129884958267212, -0.0415368489921093, -0.021276459097862244, -0.037051036953926086, -0.003961151000112295, -0.012195427902042866, 0.05888688191771507, 0.17549072206020355, 0.13628509640693665, -0.02320612221956253, -0.0005762948421761394, -0.07287558168172836, 0.0054517025128006935, -0.1596335768699646, 0.039476871490478516, 0.010477504692971706, -0.09216652065515518, 0.035112589597702026, 0.15273672342300415, -0.17437252402305603, -0.10254407674074173, -0.0665324479341507, -0.10195345431566238, -0.05854901298880577, -0.02167181484401226, 0.20152230560779572, 0.0725846216082573, -0.029463371261954308, -0.05362895131111145, -0.05502769351005554, 0.04964980110526085, 0.01344746258109808, -0.13519147038459778, -0.06503477692604065, 0.011956116184592247, -0.09078998118638992, 0.11130670458078384, -0.011851485818624496, 0.07664316147565842, 0.028516508638858795, 0.06955713033676147, -0.0758107528090477, 0.07879823446273804, -0.010527635924518108, -0.20249006152153015, -0.034744180738925934, 0.09050950407981873, -0.09081694483757019, -0.0027074019890278578, -0.032710831612348557, 0.03550563380122185, -0.03906232863664627, 0.006731655448675156, 0.022800317034125328, -0.11326833814382553, 0.027772346511483192, -0.14535017311573029, 0.05826260894536972, -0.02095666155219078, -0.00907511729747057, -0.02822152152657509, -0.10655737668275833, -0.018665220588445663, 0.10848913341760635, -0.013801720924675465, 0.045932359993457794, -0.14252617955207825, -0.041169095784425735, 0.08111895620822906, 0.06676296144723892, -0.09743200987577438, 0.05561665818095207, -0.11811143159866333, -0.02023584581911564, -0.12130608409643173, -0.04813636839389801, 0.05508257821202278, -0.026869889348745346, 0.008381340652704239, 0.01587291993200779, -0.005841925274580717, 0.03725558519363403, -0.15612387657165527, -0.0928182303905487 ]
null
null
transformers
This is for (private) DEMO only.
{"language": "ja", "tags": ["audio", "audio-classification", "speech", "speech-emotion-recognition"], "datasets": ["jtes"]}
audio-classification
Bagus/wav2vec2-xlsr-japanese-speech-emotion-recognition
[ "transformers", "pytorch", "safetensors", "wav2vec2", "audio-classification", "audio", "speech", "speech-emotion-recognition", "ja", "dataset:jtes", "endpoints_compatible", "has_space", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[ "ja" ]
TAGS #transformers #pytorch #safetensors #wav2vec2 #audio-classification #audio #speech #speech-emotion-recognition #ja #dataset-jtes #endpoints_compatible #has_space #region-us
This is for (private) DEMO only.
[]
[ "TAGS\n#transformers #pytorch #safetensors #wav2vec2 #audio-classification #audio #speech #speech-emotion-recognition #ja #dataset-jtes #endpoints_compatible #has_space #region-us \n" ]
[ 66 ]
[ "passage: TAGS\n#transformers #pytorch #safetensors #wav2vec2 #audio-classification #audio #speech #speech-emotion-recognition #ja #dataset-jtes #endpoints_compatible #has_space #region-us \n" ]
[ -0.11270296573638916, 0.10602933168411255, -0.006220296025276184, -0.006839474197477102, 0.07744733989238739, -0.031652141362428665, 0.024517465382814407, 0.0914294645190239, 0.017148097977042198, 0.054602086544036865, 0.030701082199811935, 0.14332814514636993, -0.00839153677225113, 0.04091492295265198, -0.10599809885025024, -0.24101264774799347, 0.06722795218229294, 0.06565529853105545, 0.0977124571800232, 0.10245812684297562, 0.06341348588466644, -0.12109322100877762, 0.0190244372934103, 0.015369391068816185, -0.10914099216461182, 0.0038033395539969206, 0.04031399264931679, -0.11552369594573975, 0.10223956406116486, -0.018872659653425217, 0.09136159718036652, 0.0637417584657669, 0.0004608154413290322, -0.17784127593040466, 0.03601119667291641, 0.014201542362570763, -0.001446774578653276, 0.02767251804471016, 0.06477689743041992, -0.09677218645811081, 0.0004959302023053169, -0.036912258714437485, -0.035030193626880646, 0.053711291402578354, -0.08308958262205124, -0.25424808263778687, 0.008315334096550941, 0.06350499391555786, 0.026057252660393715, 0.07092568278312683, -0.05588650703430176, 0.14791209995746613, -0.10962846875190735, 0.11506080627441406, 0.22092027962207794, -0.19547919929027557, 0.0027279460337013006, 0.02153242565691471, 0.15051808953285217, 0.02139413356781006, -0.09624984860420227, 0.07140685617923737, 0.04654762148857117, 0.01768345758318901, 0.01928841695189476, -0.08096946030855179, -0.17220023274421692, -0.0011304387589916587, -0.10136174410581589, -0.03412467986345291, 0.23603682219982147, -0.0008918323437683284, 0.041507914662361145, -0.09535469114780426, -0.05193718150258064, -0.08739030361175537, -0.04521322622895241, -0.0009727627038955688, 0.0021125946659594774, 0.003795386990532279, -0.0325741283595562, 0.06613542139530182, -0.11621705442667007, -0.022256873548030853, -0.09387999027967453, 0.2013985812664032, -0.03116283006966114, 0.03865431249141693, -0.12168415635824203, -0.001954813487827778, -0.004837609827518463, -0.10829099267721176, 0.03305636718869209, -0.050489481538534164, -0.00018010394705925137, -0.0026981045957654715, -0.06040539592504501, 0.0032891291193664074, 0.09714379161596298, -0.04188944771885872, -0.06443575769662857, 0.02822769060730934, -0.051147717982530594, 0.08109406381845474, 0.12370959669351578, 0.062498923391103745, -0.043799061328172684, -0.05786578357219696, -0.011079107411205769, -0.03600470349192619, 0.03208446502685547, -0.025092165917158127, -0.08939141780138016, -0.03507654741406441, 0.04367075487971306, 0.05389358848333359, 0.02416183240711689, 0.039507605135440826, -0.10558987408876419, 0.011973721906542778, -0.010114941745996475, -0.06135369837284088, 0.02201768197119236, 0.07394056767225266, 0.0935327559709549, 0.1239641085267067, -0.024483919143676758, 0.04132948815822601, -0.03687802329659462, 0.04443919286131859, 0.017442218959331512, 0.0073027885518968105, 0.032364871352910995, -0.057059455662965775, 0.0729873850941658, -0.11401490122079849, 0.08814165741205215, -0.15046918392181396, -0.03314399719238281, -0.035629868507385254, 0.0008215588168241084, 0.028380507603287697, -0.06460527330636978, -0.010852660052478313, -0.05682859942317009, 0.0493403859436512, -0.1090635135769844, -0.05849451944231987, -0.08711103349924088, 0.08411470055580139, -0.04677252098917961, 0.14125148952007294, -0.06655975431203842, 0.07925614714622498, -0.07164836674928665, 0.025369657203555107, -0.03974464535713196, 0.04107017442584038, -0.06821611523628235, 0.07523787021636963, 0.00216874573379755, -0.07411050796508789, -0.10308557748794556, 0.06370219588279724, -0.06478443741798401, 0.125776007771492, -0.2144540548324585, -0.12994250655174255, 0.16693201661109924, -0.10698282718658447, -0.10781500488519669, 0.12604068219661713, 0.02640812285244465, -0.03857707977294922, 0.0938989669084549, 0.3367163836956024, 0.03239540755748749, -0.13087356090545654, -0.07603473216295242, 0.09600695222616196, -0.06100727617740631, -0.018874645233154297, 0.043083664029836655, 0.048218850046396255, 0.04479732736945152, -0.0018986627692356706, 0.1010529026389122, 0.05034507438540459, -0.06334543973207474, -0.05525000020861626, -0.038835473358631134, -0.07362034916877747, 0.08018258213996887, 0.02458120882511139, -0.0039768340066075325, -0.06220119446516037, -0.008353062905371189, 0.006415619049221277, 0.05853892117738724, -0.025843100622296333, 0.06449202448129654, -0.08182929456233978, 0.16206122934818268, -0.04416760057210922, -0.02320719137787819, -0.1854354739189148, 0.15334844589233398, -0.04511348158121109, 0.07452747225761414, 0.01315395813435316, 0.14400796592235565, 0.07440023124217987, -0.07645698636770248, -0.00888158194720745, -0.09157297760248184, 0.10429500788450241, 0.07615271955728531, 0.006191895809024572, -0.197244331240654, 0.04187926650047302, -0.10334283113479614, -0.03584117442369461, -0.023866921663284302, -0.034397192299366, 0.15435151755809784, 0.1347579061985016, 0.027910908684134483, 0.0050665149465203285, 0.07590857148170471, 0.03741008788347244, -0.0036533239763230085, 0.006032077595591545, 0.08037453889846802, 0.0017675883136689663, -0.010417687706649303, 0.22961485385894775, -0.17975719273090363, 0.3324378728866577, 0.21913501620292664, -0.2302214503288269, 0.06195817142724991, 0.09595374017953873, -0.024150395765900612, 0.03810039907693863, 0.02145889587700367, -0.008111688308417797, 0.13596752285957336, -0.052981872111558914, 0.09044626355171204, -0.06586485356092453, -0.029606012627482414, 0.06069406867027283, -0.052746303379535675, -0.05893347039818764, 0.04982540383934975, -0.11109282821416855, -0.09032755345106125, 0.14782197773456573, 0.24150791764259338, -0.03887022286653519, 0.2404562532901764, -0.031564533710479736, -0.012405648827552795, 0.030684882774949074, -0.06727388501167297, -0.09262825548648834, 0.10575810074806213, -0.29384517669677734, -0.06052429974079132, 0.066398024559021, -0.008189702406525612, 0.03963093087077141, -0.13255523145198822, -0.01901434361934662, 0.01800677739083767, -0.018064921721816063, -0.12511563301086426, 0.07375971972942352, 0.0340338759124279, 0.09677755832672119, -0.05849318578839302, -0.0699910968542099, 0.03524836152791977, -0.03340265527367592, -0.08269374072551727, 0.07734880596399307, -0.1369011104106903, -0.31376102566719055, -0.037534356117248535, -0.09087128192186356, 0.04467187076807022, 0.040639400482177734, 0.11107218265533447, -0.13354220986366272, -0.007385930046439171, 0.015533547848463058, 0.05551601201295853, -0.054829537868499756, 0.04497818276286125, 0.037666913121938705, 0.04332365095615387, -0.007867060601711273, -0.08434305340051651, -0.013265714049339294, -0.050181638449430466, 0.053177621215581894, 0.09167513996362686, 0.018917731940746307, 0.05605410411953926, 0.20623373985290527, 0.06329374015331268, 0.018833359703421593, -0.03665723651647568, 0.1855057030916214, -0.1495874971151352, -0.01724831946194172, 0.15511874854564667, -0.11539467424154282, 0.0023608726914972067, 0.22901450097560883, 0.04216373339295387, -0.041651200503110886, -0.029810378327965736, -0.0272227693349123, -0.07185062766075134, -0.20947034657001495, -0.13157902657985687, -0.11055386066436768, 0.06880486011505127, -0.05750264972448349, 0.04631957411766052, 0.04906269535422325, 0.009835895150899887, -0.008950729854404926, -0.11488639563322067, 0.005590343382209539, -0.0046798791736364365, 0.12727972865104675, -0.08336364477872849, 0.09507409483194351, -0.03017508238554001, -0.10283008217811584, 0.06949707120656967, 0.09278545528650284, 0.0822567343711853, 0.0989123210310936, -0.0017003703396767378, 0.058083515614271164, 0.09678774327039719, 0.15444476902484894, 0.07881202548742294, 0.016225535422563553, -0.024131931364536285, 0.0002954968367703259, -0.046695299446582794, -0.04018883779644966, 0.028982510790228844, 0.2206224650144577, -0.022679539397358894, -0.061766088008880615, -0.16681700944900513, 0.05892680585384369, 0.09397765249013901, 0.11416337639093399, -0.18123801052570343, 0.01627398282289505, 0.09691541641950607, -0.01580953039228916, -0.07670120894908905, 0.11655133217573166, 0.14590802788734436, -0.059415366500616074, 0.08687577396631241, 0.06607227027416229, 0.07343081384897232, -0.09038946777582169, 0.08419223874807358, -0.09105128794908524, -0.17459416389465332, 0.022142423316836357, 0.012710100039839745, -0.14490659534931183, 0.23712043464183807, 0.01604468561708927, 0.003836429212242365, -0.02104983665049076, -0.017249036580324173, 0.044698186218738556, 0.14394640922546387, 0.18171511590480804, 0.030763905495405197, -0.05813004449009895, -0.12086541205644608, -0.009269192814826965, 0.01691887155175209, 0.10739005357027054, 0.08718518167734146, -0.03934740275144577, 0.001259934506379068, -0.039731815457344055, 0.018688755109906197, -0.012133922427892685, -0.09694600850343704, -0.075246661901474, 0.02219654992222786, 0.23473325371742249, 0.09092589467763901, -0.0012258958304300904, -0.0965021550655365, -0.21862833201885223, 0.09013698250055313, -0.05571763589978218, 0.00002702517122088466, -0.07499222457408905, -0.14344297349452972, 0.096974678337574, 0.0016119382344186306, 0.06181599199771881, 0.004500837530940771, 0.06858614087104797, -0.07840317487716675, -0.12628640234470367, 0.1313127726316452, -0.14401787519454956, -0.053420186042785645, -0.065425343811512, 0.24792622029781342, -0.016370663419365883, 0.07698872685432434, 0.05255485326051712, 0.047380879521369934, -0.06349299103021622, -0.03559686988592148, 0.09893913567066193, 0.040172774344682693, -0.02206122875213623, 0.06966706365346909, 0.029595382511615753, -0.2576282024383545, 0.0056257592514157295, 0.019952712580561638, 0.20667120814323425, 0.14627890288829803, -0.05700656399130821, 0.17857451736927032, 0.15803152322769165, -0.02246239222586155, -0.35894086956977844, -0.03376458212733269, -0.03251649811863899, 0.03138319030404091, -0.0054464079439640045, -0.06805469840765, 0.10321304947137833, -0.11316945403814316, -0.08603707700967789, 0.010494433343410492, -0.13548104465007782, -0.06868159770965576, 0.23322440683841705, -0.04701105132699013, 0.3289949297904968, -0.11016435921192169, -0.05452483892440796, -0.024384375661611557, -0.1483326256275177, 0.12473826855421066, -0.22708889842033386, 0.07389279454946518, 0.04195666313171387, 0.057195063680410385, 0.052509233355522156, -0.03751311078667641, 0.09766948968172073, 0.06319258362054825, -0.004942128900438547, -0.04625637084245682, -0.043017465621232986, 0.03606649488210678, -0.003971474710851908, -0.029249409213662148, -0.016341667622327805, 0.022443557158112526, -0.1314261108636856, -0.032200198620557785, -0.11240016669034958, 0.07280552387237549, 0.033440202474594116, -0.03048631176352501, -0.005352307111024857, -0.024544548243284225, 0.0310564786195755, 0.01194741204380989, 0.18007473647594452, -0.0935639813542366, 0.06899010390043259, 0.1867343634366989, 0.18732194602489471, -0.10883285850286484, -0.06790398806333542, -0.011864844709634781, -0.08250715583562851, 0.11079318076372147, -0.09646323323249817, 0.08155877888202667, 0.08635881543159485, 0.02162337675690651, 0.031968891620635986, 0.09523209184408188, -0.03716204687952995, 0.011673439294099808, 0.12294325232505798, -0.13004907965660095, -0.049813058227300644, -0.04646601155400276, -0.017579399049282074, 0.05131854861974716, 0.0934220552444458, 0.11725342273712158, 0.0026913064066320658, -0.020549016073346138, -0.032752979546785355, -0.022031445056200027, -0.11992081254720688, 0.10925416648387909, 0.04551520198583603, 0.04692968353629112, -0.1536213904619217, 0.030148595571517944, -0.055383361876010895, -0.2335267812013626, 0.0010866080410778522, -0.009647204540669918, -0.08702879399061203, -0.1256670355796814, -0.0017967172898352146, 0.030486296862363815, 0.05357887223362923, -0.08762403577566147, 0.027303891256451607, -0.1975918859243393, 0.03828747197985649, 0.22962695360183716, 0.06024355813860893, 0.08966854959726334, -0.07107210904359818, -0.020244617015123367, -0.009374878369271755, 0.06777697056531906, 0.011610022746026516, 0.015140792354941368, -0.2132325917482376, 0.015648968517780304, -0.016834501177072525, 0.11408652365207672, -0.11133306473493576, -0.03951117396354675, -0.1260271519422531, 0.05699469894170761, -0.06551158428192139, -0.008952774107456207, -0.09171450883150101, -0.022301172837615013, 0.003507402492687106, -0.06041152775287628, -0.04335218667984009, 0.008068294264376163, -0.09302147477865219, 0.076919324696064, 0.015452342107892036, 0.06673962622880936, -0.08631332963705063, -0.03677903115749359, 0.04445534199476242, -0.055895667523145676, 0.11930498480796814, 0.15732772648334503, -0.11628902703523636, 0.06670127063989639, -0.22509963810443878, -0.18719220161437988, 0.20668095350265503, 0.005535659845918417, 0.015333202667534351, -0.002717333845794201, 0.0029649632051587105, 0.08253546804189682, 0.025353193283081055, 0.022010022774338722, 0.05621495470404625, -0.026717308908700943, 0.04936732351779938, -0.06467798352241516, -0.08298379182815552, -0.003218643367290497, -0.06097709760069847, 0.13813360035419464, 0.00026895973132923245, 0.10191740095615387, -0.03302992880344391, 0.031740445643663406, -0.08561313152313232, 0.046373795717954636, -0.04320096969604492, -0.1677849143743515, -0.1498691886663437, -0.02572425827383995, 0.048780374228954315, -0.051513366401195526, 0.2239709049463272, 0.023744354024529457, -0.09110414981842041, 0.00769805908203125, 0.0507497675716877, -0.06692764163017273, 0.034094035625457764, 0.24940994381904602, 0.06136932224035263, -0.08238615840673447, -0.06927222013473511, -0.00778884207829833, 0.05942307785153389, 0.05147755891084671, -0.03939121216535568, 0.1767158955335617, 0.1293456256389618, 0.08168279379606247, 0.08347233384847641, 0.015695471316576004, -0.13055196404457092, -0.047746673226356506, -0.10482829809188843, 0.08119188249111176, -0.009286348707973957, 0.09430046379566193, 0.14023692905902863, 0.0032754442654550076, 0.09189353883266449, -0.06016647443175316, -0.00740510318428278, -0.17383651435375214, -0.09114083647727966, -0.083774633705616, -0.12141359597444534, -0.003650184255093336, -0.11008050292730331, 0.04191140457987785, -0.006167449522763491, 0.049373526126146317, -0.00498806219547987, 0.1232113316655159, 0.007396381348371506, -0.03059726022183895, 0.1392994374036789, 0.004615556448698044, -0.023377640172839165, -0.029071280732750893, 0.007923849858343601, 0.061300624161958694, -0.02485145628452301, -0.00847382377833128, -0.019123097881674767, -0.13057957589626312, 0.019079947844147682, -0.13496838510036469, -0.108759805560112, -0.01862969994544983, 0.003246207255870104, 0.014330342411994934, 0.10991013795137405, 0.030867062509059906, -0.030504805967211723, 0.0430930070579052, 0.15940190851688385, -0.09209313243627548, -0.06888632476329803, -0.037036340683698654, 0.048763569444417953, -0.04286591708660126, 0.1019931212067604, -0.023015279322862625, -0.03388841822743416, -0.07951700687408447, 0.19038861989974976, 0.2664746344089508, -0.0464429147541523, 0.0837283581495285, -0.009562921710312366, 0.03459380939602852, -0.05204533040523529, -0.008470633067190647, 0.1536521166563034, 0.2186596393585205, -0.02938714064657688, -0.0313374362885952, -0.06542807072401047, -0.03315692022442818, -0.012302626855671406, 0.022075742483139038, 0.023151885718107224, -0.1026698648929596, -0.018371829763054848, 0.10446693748235703, -0.23465172946453094, 0.06059509143233299, -0.0243955310434103, -0.24301466345787048, -0.05757909640669823, -0.026576945558190346, 0.17099730670452118, 0.1312938928604126, 0.020644277334213257, -0.03541381284594536, -0.1046992614865303, 0.08229964226484299, -0.005095959175378084, -0.17725440859794617, 0.06712581217288971, 0.00208056578412652, -0.1214301809668541, 0.0003517436271067709, -0.04992803931236267, 0.0840601772069931, 0.013511372730135918, 0.09159569442272186, -0.016078228130936623, 0.09083806723356247, 0.026982296258211136, -0.19325627386569977, -0.04563300311565399, 0.21875430643558502, -0.030744602903723717, 0.06122833117842674, 0.0741628110408783, -0.18923737108707428, 0.042810384184122086, -0.040472760796546936, -0.0513298399746418, -0.05291641503572464, 0.019935578107833862, -0.028037061914801598, 0.027582619339227676, -0.045035749673843384, -0.013506593182682991, -0.03830447047948837, -0.020834162831306458, -0.01423038449138403, 0.039516858756542206, -0.017825668677687645, -0.038428258150815964, -0.129727303981781, -0.03008405491709709, -0.007805239409208298, 0.014879058115184307, -0.1460801213979721, -0.03442549705505371, -0.06010695919394493, 0.02636277675628662, -0.11667119711637497, -0.010432940907776356, 0.03562517464160919, -0.0031093712896108627, 0.013552272692322731, -0.06534168124198914, 0.113064706325531, 0.1275205910205841, -0.09321819990873337, -0.0687389224767685 ]
null
null
transformers
Harry Potter DialoGPT Model
{"tags": ["conversational"]}
text-generation
BalajiSathesh/DialoGPT-small-harrypotter
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Harry Potter DialoGPT Model
[]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 51 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.009697278961539268, 0.03208012506365776, -0.007204889785498381, 0.004809224978089333, 0.16726240515708923, 0.014898733235895634, 0.09765533357858658, 0.13672804832458496, -0.007841327227652073, -0.031050153076648712, 0.14490588009357452, 0.20411323010921478, -0.006439372431486845, 0.0661218985915184, -0.07572533935308456, -0.2683109939098358, 0.05759621039032936, 0.046649303287267685, 0.016515716910362244, 0.1200079694390297, 0.08573378622531891, -0.05473608896136284, 0.08714032918214798, -0.014583407901227474, -0.150366872549057, 0.017733458429574966, 0.043394338339567184, -0.12260226160287857, 0.11910516023635864, 0.05462685227394104, 0.07063519209623337, 0.014929565601050854, -0.07541623711585999, -0.1631229966878891, 0.03031250834465027, 0.01425902172923088, -0.0594632662832737, 0.04757995903491974, 0.059961482882499695, -0.10165371745824814, 0.10819483548402786, 0.09530027210712433, -0.013078106567263603, 0.06798283755779266, -0.16849711537361145, -0.020869607105851173, -0.01446688175201416, 0.009899779222905636, 0.05550243332982063, 0.09964893013238907, -0.03413357585668564, 0.10497362166643143, -0.09214533120393753, 0.11017382889986038, 0.10932035744190216, -0.32057443261146545, -0.005767723545432091, 0.09167823940515518, 0.039358653128147125, 0.07352814823389053, -0.04467793554067612, 0.06258884817361832, 0.018015462905168533, 0.017986174672842026, -0.014015024527907372, -0.07283061742782593, -0.11612214148044586, 0.04717336222529411, -0.08668071031570435, -0.059868961572647095, 0.2244078367948532, -0.05464440956711769, 0.06881742179393768, -0.05281897634267807, -0.10522868484258652, -0.04308144748210907, -0.029833965003490448, 0.00475557055324316, -0.07660607248544693, 0.08692064881324768, 0.00869679357856512, -0.09547875821590424, -0.1376667022705078, -0.02496783249080181, -0.1776352822780609, 0.16140350699424744, 0.02465328387916088, 0.05232657864689827, -0.2027255892753601, 0.09623090922832489, 0.017906051129102707, -0.08045592904090881, 0.022091427817940712, -0.10046248883008957, 0.029131146147847176, 0.013760408386588097, -0.04754498973488808, -0.061387211084365845, 0.0843690037727356, 0.11199145019054413, -0.01731434464454651, 0.025486016646027565, -0.039331406354904175, 0.08100687712430954, 0.03553595021367073, 0.09077847748994827, 0.007288969587534666, -0.028338588774204254, 0.025842782109975815, -0.13719046115875244, -0.003647835226729512, -0.07116208970546722, -0.16572439670562744, -0.021088803187012672, 0.02994808368384838, 0.08289173990488052, 0.015449047088623047, 0.11682453751564026, -0.03272046521306038, -0.025152435526251793, 0.03602350503206253, -0.047656361013650894, -0.012649794109165668, 0.016648368909955025, 0.013163427822291851, 0.12399329990148544, -0.0022096503525972366, 0.03235051408410072, -0.13653022050857544, 0.031423524022102356, -0.06793295592069626, -0.003740974934771657, -0.03486552834510803, -0.040637075901031494, 0.009043924510478973, -0.06862333416938782, 0.003486064961180091, -0.15030112862586975, -0.15063877403736115, 0.007587034720927477, -0.007836631499230862, -0.04107699543237686, -0.06370922178030014, -0.06952770054340363, -0.013550350442528725, 0.04251532256603241, -0.07093454152345657, -0.011352915316820145, -0.06403283774852753, 0.11004766076803207, -0.03197755664587021, 0.07921615242958069, -0.11953279376029968, 0.08390819281339645, -0.11260783672332764, -0.02386913076043129, -0.060801517218351364, 0.09317506104707718, -0.0006014376995153725, 0.09549830108880997, -0.006563255097717047, -0.017931854352355003, -0.07981178909540176, 0.06445012241601944, -0.042872510850429535, 0.21701598167419434, -0.0615808479487896, -0.11181682348251343, 0.28781595826148987, -0.052628401666879654, -0.1370542049407959, 0.11647392809391022, 0.008682746440172195, 0.05777018144726753, 0.10703510791063309, 0.19733482599258423, -0.015276194550096989, 0.004040541127324104, 0.09471915662288666, 0.11263324320316315, -0.11276852339506149, -0.033160366117954254, 0.013019153848290443, -0.04081077128648758, -0.10867965966463089, 0.04689536616206169, 0.09810488671064377, 0.07090286910533905, -0.04786505550146103, -0.03377414867281914, -0.01366397924721241, 0.0052589005790650845, 0.08885077387094498, -0.007157256826758385, 0.10962837189435959, -0.05819983780384064, -0.03796621412038803, -0.029282379895448685, -0.012126247398555279, -0.03951939567923546, 0.03137664496898651, -0.043376367539167404, 0.10821941494941711, -0.011204327456653118, 0.06364280730485916, -0.16185984015464783, -0.07691477984189987, -0.017002692446112633, 0.1581239402294159, 0.024538565427064896, 0.09859629720449448, 0.0552486926317215, -0.040398042649030685, -0.0012767292791977525, 0.012792680412530899, 0.15581141412258148, -0.022091681137681007, -0.065607450902462, -0.052166227251291275, 0.08642971515655518, -0.05641226842999458, 0.04504093527793884, -0.05937713757157326, 0.012367865070700645, 0.05064384639263153, 0.10342344641685486, -0.00018274025933351368, 0.03323284164071083, -0.008164864964783192, 0.002145637758076191, -0.058205123990774155, 0.007405933458358049, 0.10799351334571838, 0.00036868182360194623, -0.07365862280130386, 0.22074243426322937, -0.17796069383621216, 0.1765957772731781, 0.1893044263124466, -0.299345999956131, 0.017949223518371582, -0.10759581625461578, -0.04561871662735939, 0.014407722279429436, 0.05567655712366104, -0.0454222597181797, 0.1703362911939621, -0.009871348738670349, 0.18874616920948029, -0.04946064203977585, -0.04464937001466751, -0.0200483538210392, -0.05118836089968681, -0.0024189651012420654, 0.07781197130680084, 0.10685696452856064, -0.13992026448249817, 0.1964332014322281, 0.1621224284172058, 0.048237916082143784, 0.19945049285888672, 0.015346456319093704, -0.011589210480451584, 0.0909530371427536, 0.005220826715230942, -0.058739423751831055, -0.07409929484128952, -0.2594851851463318, -0.030033592134714127, 0.07992640137672424, 0.0422382652759552, 0.1212305948138237, -0.11349532753229141, -0.038956157863140106, -0.01763172075152397, -0.023146281018853188, 0.021672505885362625, 0.0914369598031044, 0.06075398623943329, 0.13201528787612915, -0.001710098935291171, -0.007300339173525572, 0.10524573177099228, 0.01783694699406624, -0.09354141354560852, 0.18308524787425995, -0.13652534782886505, -0.37097251415252686, -0.13911493122577667, -0.18057456612586975, -0.05449081212282181, 0.05712554603815079, 0.11679314076900482, -0.12011238187551498, -0.018752124160528183, 0.01578843593597412, 0.10931742936372757, -0.08449502289295197, 0.0021454424131661654, -0.06880278885364532, 0.0321490578353405, -0.10310184955596924, -0.09194442629814148, -0.055416494607925415, -0.031392451375722885, -0.08001253753900528, 0.1423761546611786, -0.10777941346168518, 0.04476889222860336, 0.20262959599494934, 0.04653622955083847, 0.05625178664922714, -0.044105201959609985, 0.19377262890338898, -0.11264272034168243, -0.01661740615963936, 0.19215328991413116, -0.048360925167798996, 0.07476246356964111, 0.1232115849852562, -0.006348740309476852, -0.08765771239995956, 0.03011748194694519, -0.02085109055042267, -0.07988511025905609, -0.23219464719295502, -0.13938382267951965, -0.12429051846265793, 0.09477275609970093, 0.028005298227071762, 0.056365787982940674, 0.17219258844852448, 0.06577219814062119, -0.038416244089603424, 0.006410336587578058, 0.02959546446800232, 0.08237514644861221, 0.23417828977108002, -0.06035616248846054, 0.1364797055721283, -0.03420931473374367, -0.14982740581035614, 0.08169995993375778, 0.0713929831981659, 0.10213395953178406, 0.06678459793329239, 0.0804823637008667, 0.0149586396291852, 0.06188136339187622, 0.1311223804950714, 0.08191446959972382, 0.019586285576224327, -0.02480296604335308, -0.03388110175728798, -0.025523077696561813, -0.05937909707427025, 0.040128443390131, 0.06589099019765854, -0.16763372719287872, -0.039227183908224106, -0.09338314831256866, 0.09657008945941925, 0.0873042419552803, 0.06609832495450974, -0.1842060089111328, -0.008006223477423191, 0.08488986641168594, -0.03854905813932419, -0.13727426528930664, 0.09535189718008041, 0.01523482333868742, -0.15144726634025574, 0.03139317408204079, -0.04061909019947052, 0.12188644707202911, -0.07804752141237259, 0.09809603542089462, -0.08108244836330414, -0.07448557764291763, 0.02123199962079525, 0.1261177361011505, -0.30527687072753906, 0.20240111649036407, -0.0024993624538183212, -0.06486981362104416, -0.1243603527545929, -0.0032166161108762026, 0.002410882618278265, 0.07357452809810638, 0.10519039630889893, -0.007196315098553896, 0.001897757756523788, -0.06300821900367737, -0.01829923689365387, 0.032471053302288055, 0.13080233335494995, -0.0401318334043026, -0.021158374845981598, -0.050194524228572845, -0.001653497340157628, -0.03173094615340233, -0.06934895366430283, 0.02002747356891632, -0.19509181380271912, 0.08751901984214783, 0.04166261479258537, 0.09648149460554123, 0.029994789510965347, 0.004265148192644119, -0.09651939570903778, 0.24698667228221893, -0.07148019969463348, -0.10072879493236542, -0.10919588059186935, -0.046813901513814926, 0.03569883480668068, -0.05628936365246773, 0.04309194162487984, -0.0788632407784462, 0.028997479006648064, -0.06352769583463669, -0.19235502183437347, 0.12410202622413635, -0.09027006477117538, -0.04412810131907463, -0.02371402643620968, 0.2110891044139862, -0.05598580464720726, 0.010335659608244896, 0.02930437959730625, 0.01208863127976656, -0.11645778268575668, -0.09678568691015244, 0.031018631532788277, -0.007351789623498917, 0.050603240728378296, 0.041841957718133926, -0.05915454775094986, -0.017138581722974777, -0.052199993282556534, -0.022926922887563705, 0.3496883809566498, 0.14231905341148376, -0.043836336582899094, 0.19347235560417175, 0.12347975373268127, -0.07452994585037231, -0.3159443140029907, -0.1066238060593605, -0.10937739163637161, -0.04680149629712105, -0.07012093812227249, -0.2002030611038208, 0.06474938243627548, 0.00662544509395957, -0.013415241613984108, 0.12749312818050385, -0.2561831772327423, -0.07571036368608475, 0.15906259417533875, -0.017980827018618584, 0.3745945692062378, -0.1168576180934906, -0.10926306992769241, -0.03950892388820648, -0.14175476133823395, 0.16968177258968353, -0.01989765651524067, 0.11221715062856674, -0.009765521623194218, 0.14388824999332428, 0.05548359826207161, -0.023479344323277473, 0.08544106781482697, 0.004999885335564613, -0.03290518373250961, -0.10304180532693863, -0.05676887184381485, 0.007092386484146118, 0.02477436140179634, 0.018026655539870262, -0.041834570467472076, 0.02227151393890381, -0.11731979995965958, -0.04657655209302902, -0.08982590585947037, 0.04431166127324104, 0.03899754583835602, -0.07325074821710587, -0.002380647463724017, -0.07165111601352692, -0.012272949330508709, 0.022334342822432518, 0.20356793701648712, -0.08029330521821976, 0.16448934376239777, 0.09239562600851059, 0.12419285625219345, -0.14376309514045715, -0.00019283240544609725, -0.0762530043721199, -0.05611240118741989, 0.07737895101308823, -0.09433035552501678, 0.058893077075481415, 0.10901971161365509, -0.04567738622426987, 0.08828683942556381, 0.10377411544322968, 0.008936077356338501, 0.003213887568563223, 0.10916902124881744, -0.2667325437068939, -0.0296600554138422, -0.07532413303852081, 0.000883326749317348, 0.09092561900615692, 0.08562852442264557, 0.18840822577476501, 0.025361526757478714, -0.04293036088347435, -0.002770674182102084, 0.028597986325621605, -0.039021048694849014, 0.051667019724845886, 0.001123449532315135, 0.01947369985282421, -0.1530752182006836, 0.072522833943367, 0.01490565575659275, -0.15215420722961426, 0.021316176280379295, 0.16572684049606323, -0.11656328290700912, -0.1283872276544571, -0.06520111113786697, 0.08313824236392975, -0.11755692958831787, -0.01578943058848381, -0.03279297426342964, -0.13145680725574493, 0.07992171496152878, 0.12629036605358124, 0.05557859688997269, 0.0972496047616005, -0.06061713397502899, -0.020469192415475845, -0.018721895292401314, -0.014099318534135818, -0.012384648434817791, -0.007667020428925753, -0.055978111922740936, 0.0590752474963665, -0.026677248999476433, 0.1425808072090149, -0.09221141785383224, -0.1037059873342514, -0.16142144799232483, 0.0374140702188015, -0.11013076454401016, -0.08825794607400894, -0.08821134269237518, -0.050188567489385605, 0.002360827289521694, -0.019856395199894905, -0.04037635400891304, -0.05829505994915962, -0.12300454825162888, 0.0338277705013752, -0.040771447122097015, 0.024727050215005875, -0.07512269169092178, 0.015856385231018066, 0.08507686108350754, -0.03285100311040878, 0.15655414760112762, 0.1450488418340683, -0.1006515845656395, 0.10741901397705078, -0.14806775748729706, -0.09138492494821548, 0.11116421222686768, 0.015329592861235142, 0.0449691042304039, 0.09723787009716034, 0.013362943194806576, 0.0635865181684494, 0.032776717096567154, 0.05308786407113075, 0.027619892731308937, -0.11959987878799438, 0.06483134627342224, -0.03626115620136261, -0.14700546860694885, -0.049338050186634064, -0.05282869189977646, 0.01647452637553215, 0.013054544106125832, 0.09622690081596375, -0.05301849544048309, 0.10698331147432327, -0.04055701196193695, 0.0346808135509491, 0.017554637044668198, -0.1730053424835205, -0.03816922754049301, -0.08538098633289337, 0.03681723028421402, 0.014741539023816586, 0.25266793370246887, 0.030072299763560295, 0.012416383251547813, 0.032671261578798294, 0.08285367488861084, 0.03899408504366875, 0.010228337720036507, 0.17482228577136993, 0.1162426546216011, -0.06621865928173065, -0.10445023328065872, 0.0729617029428482, 0.016332454979419708, 0.01286179106682539, 0.13617953658103943, 0.008365051820874214, 0.005795429926365614, 0.08649782836437225, -0.016865963116288185, 0.009968153201043606, -0.10052056610584259, -0.13426925241947174, -0.022176474332809448, 0.05151832848787308, -0.04655967652797699, 0.11727844923734665, 0.1406494379043579, -0.01806013658642769, 0.03222079202532768, -0.021771740168333054, -0.05699979141354561, -0.1683429479598999, -0.1429590880870819, -0.06883849948644638, -0.13416796922683716, 0.00897989235818386, -0.11180389672517776, 0.05395037308335304, 0.06001098081469536, 0.06750501692295074, -0.06899319589138031, 0.10220931470394135, 0.04626858979463577, -0.11440542340278625, 0.06264589726924896, -0.0296088308095932, 0.09430401772260666, -0.02759445086121559, -0.019505485892295837, -0.09039592742919922, 0.014574515633285046, 0.011419114656746387, 0.06245238706469536, -0.04707273095846176, 0.007463190704584122, -0.14696238934993744, -0.08972041308879852, -0.0523175448179245, 0.0718572810292244, -0.050409089773893356, 0.14282815158367157, 0.00775480642914772, -0.0170906875282526, 0.039554283022880554, 0.22787313163280487, -0.07476283609867096, -0.04778539761900902, -0.05269690603017807, 0.20717895030975342, 0.02975541539490223, 0.1171872541308403, -0.022938819602131844, -0.006106364540755749, -0.0919521227478981, 0.3764844834804535, 0.30030161142349243, -0.09031439572572708, 0.011794124729931355, 0.02137952297925949, 0.04502861574292183, 0.1316293478012085, 0.1216534823179245, 0.10318691283464432, 0.3006802201271057, -0.07452366501092911, -0.04653361067175865, -0.012629742734134197, -0.023858042433857918, -0.09059546142816544, 0.1021224707365036, 0.04839762672781944, -0.06382183730602264, -0.03313443064689636, 0.0954432487487793, -0.25862133502960205, 0.1277991235256195, -0.12311873584985733, -0.17578600347042084, -0.06654827296733856, 0.009760108776390553, 0.10465722531080246, 0.015642458572983742, 0.0946015790104866, 0.007128213066607714, -0.11252258718013763, 0.06305865943431854, 0.03397420793771744, -0.22762253880500793, 0.0006893770187161863, 0.06642123311758041, -0.07006710022687912, -0.0024247700348496437, -0.026499588042497635, 0.05657242611050606, 0.0656052976846695, 0.054629553109407425, -0.00971333310008049, 0.03816632181406021, 0.0034184439573436975, -0.0585215799510479, 0.016623929142951965, 0.05121519789099693, 0.02472509816288948, -0.09763528406620026, 0.06927435845136642, -0.1574270874261856, 0.04766253009438515, -0.0030655991286039352, -0.04124255105853081, 0.006064958870410919, 0.008823691867291927, -0.06491616368293762, 0.05165379121899605, 0.07916834205389023, -0.0016257909592241049, -0.0062433634884655476, -0.057178743183612823, -0.02632102556526661, -0.027755750343203545, -0.09291748702526093, -0.10495562851428986, -0.14682936668395996, -0.11640441417694092, 0.09368976950645447, -0.01011267676949501, -0.1848134547472, 0.022154374048113823, -0.08606051653623581, 0.08319322764873505, -0.1670055389404297, 0.08040720224380493, 0.07041648775339127, 0.013038921169936657, -0.0031511052511632442, -0.02002427540719509, 0.054132770746946335, 0.086809903383255, -0.10407156497240067, -0.07400695979595184 ]
null
null
transformers
**Dataset** ToTTo is an open-domain English Table-to-Text dataset with over 120,000 training examples that proposes a controlled generation task: given a Wikipedia table, a set of highlighted table cells, page title and section title as inputs, it produces a one-sentence description summarising the key details from the inputs. This dataset can be taken from hugging face (https://huggingface.co/datasets/totto). **Model** The pre-trained Text-to-Text "t5-base" model is fine-tuned with the Table-to-Text ToTTo dataset(downstream task) for the complete train dataset split of around 120,761 examples. During the fine-tuning process for this downstream task, BertScore metric was used as an evaluation metric instead of the standard BLEU metric.
{}
text2text-generation
Barkavi/t5base_totto
[ "transformers", "pytorch", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Dataset ToTTo is an open-domain English Table-to-Text dataset with over 120,000 training examples that proposes a controlled generation task: given a Wikipedia table, a set of highlighted table cells, page title and section title as inputs, it produces a one-sentence description summarising the key details from the inputs. This dataset can be taken from hugging face (URL Model The pre-trained Text-to-Text "t5-base" model is fine-tuned with the Table-to-Text ToTTo dataset(downstream task) for the complete train dataset split of around 120,761 examples. During the fine-tuning process for this downstream task, BertScore metric was used as an evaluation metric instead of the standard BLEU metric.
[]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 48 ]
[ "passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.01584368571639061, 0.001455417019315064, -0.00658801756799221, 0.0177968367934227, 0.18000324070453644, 0.01899094320833683, 0.1102970764040947, 0.13923293352127075, -0.029492201283574104, -0.031411342322826385, 0.1258108913898468, 0.215000182390213, -0.002026807749643922, 0.09281328320503235, -0.09747900068759918, -0.26333776116371155, 0.035797640681266785, 0.06643600016832352, 0.01654808409512043, 0.13231700658798218, 0.07867445051670074, -0.06135464087128639, 0.09729219973087311, -0.03548338636755943, -0.1792060285806656, 0.056830670684576035, 0.06633275002241135, -0.14007478952407837, 0.12120860069990158, 0.05082603916525841, 0.11879663914442062, 0.03541290760040283, -0.049473561346530914, -0.12151949107646942, 0.027005361393094063, 0.036254558712244034, -0.0702618658542633, 0.05806567892432213, 0.12953147292137146, -0.09849721938371658, 0.101417675614357, 0.05899258330464363, -0.011092896573245525, 0.06627913564443588, -0.1479889154434204, 0.00348502560518682, -0.010794720612466335, 0.021929796785116196, 0.07188789546489716, 0.09843083471059799, -0.011102980934083462, 0.12869638204574585, -0.09904436022043228, 0.14108110964298248, 0.1505395472049713, -0.3118693232536316, 0.005070185288786888, 0.04695354029536247, 0.043711405247449875, 0.07051856815814972, -0.00885665975511074, 0.03730452060699463, 0.03216231241822243, 0.03300926089286804, 0.03037545457482338, -0.07585509866476059, -0.1657770276069641, 0.04152299836277962, -0.08708652853965759, -0.0615064837038517, 0.23453429341316223, -0.06484062969684601, 0.066205233335495, -0.0072464910335838795, -0.13153931498527527, -0.06954485923051834, 0.0006044790497981012, -0.0053634620271623135, -0.055698949843645096, 0.06370817124843597, 0.016068486496806145, -0.056699495762586594, -0.1428028792142868, -0.010085990652441978, -0.20030196011066437, 0.12700670957565308, 0.0098763108253479, 0.055072084069252014, -0.23676714301109314, 0.09703067690134048, 0.05306711047887802, -0.09924419224262238, 0.062389008700847626, -0.08943060785531998, 0.014000056311488152, -0.02474105730652809, -0.06801413744688034, -0.1568288505077362, 0.06617090106010437, 0.08920741081237793, 0.003721588756889105, 0.020913373678922653, -0.08072061836719513, 0.07540135085582733, 0.015950961038470268, 0.08084282279014587, -0.009468162432312965, -0.02545289881527424, 0.050731536000967026, -0.13451236486434937, -0.00853751040995121, -0.06626977026462555, -0.15002763271331787, -0.07248537242412567, 0.08783774077892303, 0.09340707957744598, 0.024510055780410767, 0.09635060280561447, -0.0332786962389946, -0.04203520715236664, 0.009204940870404243, -0.09221477061510086, -0.024659397080540657, 0.0008180328877642751, 0.006082381121814251, 0.14675909280776978, 0.02296186424791813, 0.007892758585512638, -0.17018215358257294, 0.06483504176139832, -0.0723656713962555, -0.007938898168504238, -0.029325438663363457, -0.07534618675708771, 0.023915085941553116, -0.10881388932466507, 0.008133405819535255, -0.17433200776576996, -0.16765841841697693, 0.01697476953268051, 0.007396905682981014, -0.018866462633013725, -0.043013010174036026, -0.045346371829509735, -0.03770101070404053, 0.04327743873000145, -0.07020818442106247, 0.010033725760877132, -0.04259585589170456, 0.10675564408302307, -0.0398472361266613, 0.06604313105344772, -0.1241411417722702, 0.08126731216907501, -0.12840037047863007, -0.026247713714838028, -0.06916461884975433, 0.0695473924279213, 0.03665204346179962, 0.12047890573740005, -0.03775748983025551, -0.03622163087129593, -0.07566536217927933, 0.04284011945128441, -0.01894138753414154, 0.1950719952583313, -0.0944138839840889, -0.10338135808706284, 0.24115116894245148, -0.07720785588026047, -0.16225671768188477, 0.08871348202228546, 0.01112399436533451, 0.0527188703417778, 0.09158273041248322, 0.17049458622932434, 0.044303521513938904, -0.007278476841747761, 0.0970718041062355, 0.1028069406747818, -0.11937293410301208, -0.10302774608135223, 0.002267509698867798, -0.020705346018075943, -0.11959504336118698, 0.04289879649877548, 0.09645403176546097, 0.07236005365848541, -0.05477796122431755, -0.03351253643631935, -0.04923287779092789, -0.007655630353838205, 0.1011999323964119, 0.004759210627526045, 0.13131408393383026, -0.06056664139032364, -0.016454286873340607, 0.00607975572347641, -0.029107315465807915, -0.03001979924738407, 0.04808073490858078, -0.027159245684742928, 0.11719837784767151, -0.03463059291243553, 0.043614711612463, -0.20667698979377747, -0.08273608982563019, -0.011199901811778545, 0.16215083003044128, -0.00014803845260757953, 0.09622485190629959, 0.05017630010843277, -0.026423487812280655, -0.013082671910524368, -0.020495356991887093, 0.14194169640541077, -0.008744661696255207, -0.07192710041999817, -0.055666014552116394, 0.05686090514063835, -0.056457314640283585, -0.029490424320101738, -0.06336628645658493, 0.016674358397722244, 0.022723432630300522, 0.12466312199831009, 0.024511994794011116, 0.060258712619543076, -0.019767967984080315, 0.026875387877225876, -0.09027257561683655, 0.012787343002855778, 0.10328754037618637, -0.005430325400084257, -0.06139841303229332, 0.2012202888727188, -0.18243330717086792, 0.21768754720687866, 0.1899390071630478, -0.2988763451576233, 0.0007227785536088049, -0.05822010710835457, -0.0336349755525589, 0.0059656258672475815, 0.05502323433756828, -0.03447169065475464, 0.08369144052267075, 0.0008040695101954043, 0.20492856204509735, -0.06399808824062347, -0.05498965084552765, 0.0025857435539364815, -0.05458337441086769, -0.006101091392338276, 0.058100759983062744, 0.0824960395693779, -0.17706909775733948, 0.1713913381099701, 0.20838424563407898, 0.023575296625494957, 0.17638693749904633, -0.007809492759406567, -0.04934080317616463, 0.08402703702449799, 0.006972316186875105, -0.032568447291851044, -0.10796601325273514, -0.1734510064125061, -0.016349755227565765, 0.0808510109782219, 0.0383298397064209, 0.09945333003997803, -0.11086979508399963, -0.022686339914798737, -0.005985935218632221, -0.0060494341887533665, -0.008344912901520729, 0.09246525168418884, 0.08367523550987244, 0.14147034287452698, -0.015902556478977203, -0.008576065301895142, 0.11812435835599899, 0.015694094821810722, -0.12598107755184174, 0.19254222512245178, -0.1325615793466568, -0.3495909571647644, -0.1631333827972412, -0.16477428376674652, -0.043477918952703476, 0.048511527478694916, 0.11342941224575043, -0.10511619597673416, -0.02352173998951912, -0.0007881404599174857, 0.08415862917900085, -0.07211752235889435, 0.03677205741405487, -0.08311079442501068, 0.06613492220640182, -0.06391098350286484, -0.08130958676338196, -0.04755308851599693, -0.013113722205162048, -0.0506163015961647, 0.15241199731826782, -0.13060742616653442, 0.05517926067113876, 0.20086060464382172, -0.008659793995320797, 0.05646828934550285, -0.0447811521589756, 0.1698468029499054, -0.06439661234617233, 0.014364821836352348, 0.22857394814491272, -0.06645470857620239, 0.07465333491563797, 0.13026019930839539, -0.017004651948809624, -0.06887000799179077, 0.04647034779191017, -0.03181997686624527, -0.08305076509714127, -0.27305132150650024, -0.11097732186317444, -0.12417944520711899, 0.08551718294620514, 0.060253627598285675, 0.050218936055898666, 0.1729225218296051, 0.07009439915418625, -0.011698170565068722, 0.04006649926304817, 0.008571630343794823, 0.0824633464217186, 0.19053572416305542, -0.008197087794542313, 0.13242574036121368, -0.06245150417089462, -0.11769289523363113, 0.08968717604875565, 0.05989821255207062, 0.12555105984210968, 0.04239042103290558, 0.046467121690511703, 0.009043761529028416, 0.07225343585014343, 0.13442648947238922, 0.1651877760887146, 0.034123022109270096, -0.0027062329463660717, -0.01349344477057457, -0.028245382010936737, -0.040193621069192886, 0.037729669362306595, 0.013470759615302086, -0.12968936562538147, -0.09492611140012741, -0.07551674544811249, 0.07577058672904968, 0.12997229397296906, 0.0744793638586998, -0.240804523229599, 0.012753864750266075, 0.06319929659366608, -0.046554870903491974, -0.11563625931739807, 0.08299679309129715, -0.003961589653044939, -0.13062408566474915, 0.06372487545013428, -0.05805215612053871, 0.12147562950849533, -0.028284739702939987, 0.09374229609966278, -0.03364879637956619, -0.07243189960718155, 0.018441040068864822, 0.1096850335597992, -0.33529332280158997, 0.20487374067306519, 0.0006690678419545293, -0.06490825116634369, -0.11783778667449951, -0.0044849165715277195, -0.0012578379828482866, 0.11027327179908752, 0.09952930361032486, -0.003345120931044221, -0.03474462404847145, -0.09134820103645325, -0.0031049586832523346, 0.016545293852686882, 0.14250630140304565, -0.025273242965340614, 0.0148016894236207, -0.059562280774116516, -0.021893899887800217, -0.013236827217042446, -0.013637681491672993, -0.002603176049888134, -0.1513184756040573, 0.0671682059764862, 0.020377542823553085, 0.06982939690351486, 0.01960124634206295, -0.02438407950103283, -0.06273293495178223, 0.21248282492160797, -0.06458115577697754, -0.10695376992225647, -0.12842507660388947, -0.04645165428519249, 0.05069807916879654, -0.0799480676651001, 0.05590132996439934, -0.07412241399288177, 0.026752561330795288, -0.0460817776620388, -0.2500396966934204, 0.12516821920871735, -0.08440219610929489, -0.04263054579496384, -0.039096806198358536, 0.18710920214653015, -0.09238360822200775, 0.0015196007443591952, 0.024500641971826553, -0.00008093049837043509, -0.08618257939815521, -0.05626978352665901, -0.008611418306827545, -0.01370612159371376, 0.0605587363243103, 0.04142594337463379, -0.09551963210105896, -0.06058591976761818, -0.04161534458398819, -0.0018586971564218402, 0.33403095602989197, 0.09810057282447815, -0.046292744576931, 0.17393392324447632, 0.10699018090963364, -0.08708086609840393, -0.30292707681655884, -0.07725819945335388, -0.0799851045012474, -0.026687202975153923, -0.028951935470104218, -0.16540394723415375, 0.0818692147731781, -0.0030985758639872074, 0.010349465534090996, 0.10349910706281662, -0.24794204533100128, -0.09147637337446213, 0.1472831815481186, 0.023999102413654327, 0.3351094424724579, -0.11293166130781174, -0.09755206853151321, -0.04931047186255455, -0.14179958403110504, 0.17238929867744446, -0.054965659976005554, 0.08938152343034744, -0.03220284730195999, 0.1103312224149704, 0.057209331542253494, -0.038672249764204025, 0.03753164783120155, 0.01280028186738491, 0.004070690833032131, -0.11656955629587173, -0.03721853345632553, 0.05927279219031334, -0.01239687204360962, 0.0431542843580246, -0.030626775696873665, 0.05042644962668419, -0.11610346287488937, -0.03559141978621483, -0.09718530625104904, 0.05527614802122116, 0.033209703862667084, -0.07230573892593384, 0.02535100467503071, -0.07929795235395432, 0.026679834350943565, -0.011463316157460213, 0.19073231518268585, -0.04869036376476288, 0.16655586659908295, 0.15388649702072144, 0.13591068983078003, -0.10760082304477692, 0.03837193548679352, -0.07492130249738693, -0.06832669675350189, 0.06800366938114166, -0.10502270609140396, 0.06655241549015045, 0.12330719083547592, -0.0411357618868351, 0.06374823302030563, 0.11320103704929352, 0.02062961272895336, -0.01765989325940609, 0.1385980248451233, -0.25890034437179565, 0.023826781660318375, -0.09969114512205124, -0.053946852684020996, 0.045324306935071945, 0.06959566473960876, 0.1803196519613266, 0.01996755413711071, -0.03243176266551018, -0.010976474732160568, 0.0005780249484814703, -0.04870473966002464, 0.07170422375202179, 0.021222015842795372, 0.024676067754626274, -0.1308782398700714, 0.09323612600564957, 0.032887544482946396, -0.14486797153949738, 0.019849436357617378, 0.19134655594825745, -0.1371304839849472, -0.11486499011516571, 0.01224886067211628, 0.11295973509550095, -0.15867024660110474, -0.024805627763271332, -0.06911034882068634, -0.1232672780752182, 0.09491096436977386, 0.21401236951351166, 0.05331400781869888, 0.1008497029542923, -0.046942487359046936, -0.0496317520737648, -0.04561499506235123, 0.009240290150046349, 0.012895791791379452, 0.030672000721096992, -0.097745880484581, 0.10578422248363495, -0.040116216987371445, 0.16213259100914001, -0.0917268618941307, -0.06222947686910629, -0.14983442425727844, 0.03229285031557083, -0.15211552381515503, -0.05419791117310524, -0.06350395083427429, -0.05325648933649063, -0.01421641930937767, -0.009074408560991287, -0.04453456401824951, -0.039223432540893555, -0.1178443431854248, 0.023740172386169434, -0.04184343293309212, 0.03357211872935295, -0.07360640168190002, -0.00745047302916646, 0.0597982257604599, -0.04032554477453232, 0.12730717658996582, 0.12071295082569122, -0.11999447643756866, 0.13207589089870453, -0.13698095083236694, -0.10772223025560379, 0.10667144507169724, 0.019947899505496025, 0.057580724358558655, 0.08691609650850296, 0.024122396484017372, 0.07328153401613235, 0.017016666010022163, 0.03875018656253815, 0.022994665428996086, -0.11720026284456253, 0.02915577031672001, -0.0438992902636528, -0.14189468324184418, -0.07547144591808319, -0.034607090055942535, 0.03158587962388992, 0.008018662221729755, 0.11852530390024185, -0.053823456168174744, 0.12022940069437027, -0.07060523331165314, 0.010234953835606575, 0.010690975934267044, -0.16182446479797363, -0.06637652963399887, -0.08411522209644318, 0.032836735248565674, -0.008235974237322807, 0.18409450352191925, 0.03454678878188133, 0.05747787654399872, 0.02793210744857788, 0.07954391092061996, 0.005215851124376059, 0.020794428884983063, 0.22607649862766266, 0.07304691523313522, -0.06886540353298187, -0.1103024035692215, 0.06498146802186966, 0.008004664443433285, 0.04483301192522049, 0.1749526411294937, 0.03727349266409874, -0.03907724469900131, 0.10007185488939285, -0.019638560712337494, 0.028130175545811653, -0.11450393497943878, -0.17080892622470856, -0.00801891554147005, 0.07657715678215027, -0.011517325416207314, 0.0829968973994255, 0.16146962344646454, -0.019973335787653923, 0.030916273593902588, -0.009044856764376163, -0.056131210178136826, -0.17957162857055664, -0.1594834178686142, -0.08296467363834381, -0.10509900003671646, -0.0014652428217232227, -0.10988666117191315, 0.05996263399720192, 0.05867818742990494, 0.06662456691265106, -0.06648626923561096, 0.10458429902791977, 0.06543407589197159, -0.11919818818569183, 0.07942314445972443, -0.028797946870326996, 0.08120650798082352, 0.000997701776213944, -0.009244642220437527, -0.08453210443258286, 0.008097044192254543, -0.03108268976211548, 0.04917836934328079, -0.047017112374305725, 0.02154003456234932, -0.15374121069908142, -0.1096470057964325, -0.02257644757628441, 0.05983618274331093, -0.0428229495882988, 0.12894387543201447, 0.017068613320589066, -0.030336754396557808, 0.02801426127552986, 0.22194334864616394, -0.08584439754486084, -0.08055929839611053, -0.050540681928396225, 0.2432609349489212, 0.06307961791753769, 0.08444320410490036, 0.0028589183930307627, -0.012717257253825665, -0.09089452773332596, 0.3591165244579315, 0.2667014002799988, -0.055692918598651886, 0.02277890220284462, 0.015944819897413254, 0.0347137413918972, 0.11798495799303055, 0.16456447541713715, 0.08827649056911469, 0.25443965196609497, -0.06533562391996384, -0.018368344753980637, -0.014501972123980522, 0.000018250484572490677, -0.0930345430970192, 0.13507813215255737, 0.04284169152379036, -0.08161267638206482, -0.024515492841601372, 0.10017646849155426, -0.24063174426555634, 0.14782121777534485, -0.09359890967607498, -0.16162940859794617, -0.060918986797332764, -0.0147289102897048, 0.11586485803127289, -0.0017279664753004909, 0.08164822310209274, -0.01215168833732605, -0.08752647042274475, 0.05363667756319046, 0.029349831864237785, -0.222853422164917, 0.01617160066962242, 0.05048945173621178, -0.11925873160362244, -0.024077240377664566, -0.011171307414770126, 0.04385644197463989, 0.06714760512113571, 0.07385427504777908, -0.04466511681675911, 0.046728942543268204, -0.004126311279833317, -0.011370057240128517, 0.04598446562886238, 0.06270765513181686, 0.01540715154260397, -0.09723247587680817, 0.05001823231577873, -0.1556699126958847, 0.03317674249410629, -0.01597830280661583, -0.023336609825491905, -0.002392916241660714, -0.005987333599478006, -0.04110949859023094, 0.057095758616924286, 0.1028817892074585, -0.008210273459553719, 0.012179257348179817, -0.09038258343935013, -0.0342213474214077, -0.0023632640950381756, -0.11412134766578674, -0.09127872437238693, -0.11541768163442612, -0.10260939598083496, 0.11215279996395111, -0.009719906374812126, -0.2153121829032898, 0.021527882665395737, -0.1023377850651741, 0.04012138023972511, -0.21606485545635223, 0.10211227834224701, 0.09086371958255768, 0.011539488099515438, 0.008776325732469559, -0.008126933127641678, 0.0464082807302475, 0.10446552187204361, -0.12512299418449402, -0.09291534125804901 ]
null
null
transformers
# Hello hugging face
{"language": "en", "license": "apache-2.0", "tags": ["exbert"], "datasets": ["bookcorpus", "wikipedia"]}
fill-mask
Barytes/hellohf
[ "transformers", "tf", "bert", "fill-mask", "exbert", "en", "dataset:bookcorpus", "dataset:wikipedia", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[ "en" ]
TAGS #transformers #tf #bert #fill-mask #exbert #en #dataset-bookcorpus #dataset-wikipedia #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# Hello hugging face
[ "# Hello hugging face" ]
[ "TAGS\n#transformers #tf #bert #fill-mask #exbert #en #dataset-bookcorpus #dataset-wikipedia #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Hello hugging face" ]
[ 60, 5 ]
[ "passage: TAGS\n#transformers #tf #bert #fill-mask #exbert #en #dataset-bookcorpus #dataset-wikipedia #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Hello hugging face" ]
[ -0.03498048707842827, 0.09483439475297928, -0.0023328622337430716, 0.09001455456018448, 0.08985398709774017, 0.0668070986866951, 0.09330960363149643, 0.14158964157104492, 0.08546216040849686, -0.052717749029397964, 0.15035061538219452, 0.11249526590108871, 0.04333670064806938, 0.18437567353248596, -0.03019709140062332, -0.21450074017047882, 0.06283826380968094, 0.04070024937391281, -0.04664785787463188, 0.052862498909235, 0.13263213634490967, 0.010537954047322273, 0.12973780930042267, -0.05492996424436569, -0.11526831984519958, 0.045088812708854675, 0.05783602222800255, -0.034773088991642, 0.1419190615415573, 0.11204500496387482, 0.060577910393476486, 0.030450930818915367, -0.013128050602972507, -0.14017550647258759, 0.04391400143504143, 0.017904289066791534, -0.09005430340766907, 0.031219709664583206, -0.06843455880880356, -0.006886543706059456, 0.07435579597949982, -0.01487916149199009, 0.02312878519296646, 0.06889451295137405, -0.1220182329416275, -0.17676228284835815, -0.10667063295841217, 0.05858185514807701, 0.030048688873648643, 0.030710089951753616, 0.025175606831908226, 0.19298255443572998, -0.01968579739332199, 0.08328726142644882, 0.14418086409568787, -0.2586878836154938, -0.03109661303460598, 0.0066483817063272, 0.028936609625816345, -0.09103726595640182, -0.020898375660181046, 0.05056212469935417, 0.07042984664440155, -0.003935812506824732, 0.05740966647863388, -0.041105013340711594, -0.05507034808397293, 0.026967795565724373, -0.036553140729665756, -0.03744315356016159, 0.24378252029418945, 0.04879507049918175, -0.02436155267059803, -0.004695530515164137, -0.0820266455411911, 0.07676414400339127, -0.010102016851305962, -0.03271301090717316, 0.04194820672273636, 0.06593718379735947, -0.08477817475795746, -0.048268526792526245, -0.10253841429948807, 0.046087395399808884, -0.22364646196365356, 0.05590031296014786, -0.01592385768890381, 0.05037672445178032, -0.10906482487916946, -0.02577967382967472, -0.07143689692020416, -0.060703229159116745, -0.04398695379495621, -0.06429450958967209, 0.07072853296995163, 0.016773665323853493, -0.04051431640982628, 0.05820377171039581, 0.11890745908021927, 0.2576264441013336, 0.06725955009460449, -0.02569219283759594, -0.07614433765411377, 0.04898485913872719, 0.021623874083161354, 0.04427926987409592, 0.03346247598528862, -0.20831385254859924, 0.10437696427106857, -0.12747426331043243, 0.11056472361087799, -0.04347672313451767, -0.11060244590044022, -0.023420676589012146, -0.015433805994689465, 0.023786550387740135, 0.08954755961894989, 0.09899100661277771, -0.010846786201000214, 0.04272572323679924, 0.18011027574539185, -0.032497163861989975, 0.016473835334181786, -0.028485972434282303, 0.007227752357721329, -0.022031592205166817, 0.08126077055931091, -0.022909510880708694, -0.000623355561401695, -0.0006197098991833627, -0.07898684591054916, -0.07324633747339249, 0.01405983604490757, -0.05000205710530281, 0.09888607263565063, -0.06473136693239212, 0.06352847814559937, -0.1705808788537979, -0.13718527555465698, 0.05886092409491539, 0.07811356335878372, -0.015308202244341373, -0.06391014158725739, 0.03387831524014473, 0.048208173364400864, 0.05983531102538109, -0.04550384730100632, -0.030527595430612564, -0.0612730011343956, 0.02182825282216072, -0.004747388884425163, 0.14203694462776184, -0.21351484954357147, 0.043538399040699005, -0.07787687331438065, 0.005440962966531515, -0.060814082622528076, -0.08660008013248444, -0.07952795177698135, 0.16129817068576813, -0.006590845063328743, 0.030483519658446312, 0.0062858425080776215, 0.04780574515461922, 0.027357373386621475, 0.16886167228221893, -0.15301434695720673, -0.0915234386920929, 0.282360702753067, -0.12128669768571854, -0.24821920692920685, 0.11941324919462204, 0.006397235207259655, 0.034567397087812424, 0.028681475669145584, 0.11926484853029251, -0.05137905105948448, -0.11735668778419495, -0.009283659979701042, 0.10894906520843506, -0.09206148236989975, -0.06968043744564056, 0.009476268664002419, -0.00780999381095171, -0.06693925708532333, 0.025564778596162796, 0.07624907791614532, 0.13705892860889435, -0.014772382564842701, -0.05176765099167824, -0.07909218221902847, -0.08967114984989166, 0.035909105092287064, -0.0019501058850437403, 0.0022608491126447916, -0.14592325687408447, -0.00735694682225585, -0.0922711044549942, 0.0377805233001709, 0.04889089986681938, 0.05636483430862427, -0.08531439304351807, 0.07053249329328537, 0.0555415153503418, 0.004282655660063028, -0.0196780227124691, -0.05157902464270592, -0.022624390199780464, 0.10603263229131699, 0.018028490245342255, 0.03205055743455887, 0.09264671802520752, -0.02980010025203228, -0.043104227632284164, 0.0073256283067166805, 0.14544270932674408, 0.04580345377326012, 0.03314417600631714, -0.15572576224803925, 0.09999481588602066, -0.038939766585826874, 0.09082375466823578, 0.012226720340549946, 0.020012514665722847, -0.027420740574598312, 0.10239395499229431, -0.002253076294437051, 0.036556027829647064, -0.0619540773332119, -0.047009993344545364, -0.06815218925476074, -0.044411834329366684, 0.10549202561378479, 0.057384006679058075, -0.08527961373329163, 0.18655814230442047, -0.06765041500329971, 0.2212856113910675, 0.20535984635353088, -0.11982016265392303, -0.027713658288121223, 0.022959226742386818, -0.03790199011564255, -0.01426420547068119, 0.06540697067975998, -0.01271687913686037, -0.028498275205492973, -0.0782015323638916, 0.10535120218992233, -0.037459127604961395, -0.009261133149266243, 0.023941192775964737, -0.07198042422533035, -0.07385656237602234, -0.003679273882880807, 0.05274655669927597, -0.20245298743247986, 0.16406571865081787, 0.23654378950595856, 0.023811521008610725, 0.11138500273227692, -0.018872905522584915, 0.021606504917144775, -0.02121494710445404, -0.044998057186603546, 0.015211123041808605, 0.11178230494260788, -0.19846239686012268, -0.02094413712620735, 0.0698138177394867, -0.033727239817380905, -0.0046040876768529415, -0.0699145719408989, -0.05496090278029442, 0.051308926194906235, -0.03610055521130562, -0.03235560283064842, 0.09480714797973633, -0.0743359625339508, 0.0899137482047081, 0.0115010691806674, -0.07517649978399277, 0.07497905939817429, 0.0019477272871881723, -0.09456819295883179, 0.10813617706298828, -0.15314778685569763, -0.2990606129169464, -0.07795947045087814, -0.21111959218978882, -0.03477966785430908, 0.030177220702171326, 0.09031301736831665, -0.019360987469553947, -0.0528685599565506, -0.010380703024566174, -0.052409980446100235, 0.05582734942436218, 0.05103810504078865, -0.045521996915340424, 0.005169250071048737, 0.018860137090086937, -0.11514291912317276, -0.05095396563410759, 0.06013011932373047, -0.005591615568846464, 0.11859195679426193, -0.10465927422046661, 0.11055804044008255, -0.00693561090156436, 0.005144651513546705, 0.008349527604877949, -0.06706192344427109, 0.1995505392551422, -0.040665142238140106, 0.06977208703756332, 0.09554005414247513, -0.04329054430127144, 0.03984592482447624, 0.1442032903432846, -0.008946101181209087, -0.10064776241779327, 0.05111050233244896, -0.046063512563705444, -0.11920453608036041, -0.13692983984947205, -0.07459452003240585, -0.07811437547206879, 0.17179524898529053, 0.0405350923538208, 0.029467716813087463, 0.17140553891658783, 0.05514638125896454, 0.015316510573029518, 0.0016998519422486424, -0.05401642248034477, 0.05046021565794945, 0.05806770175695419, -0.04956032335758209, 0.07534209638834, -0.08351607620716095, -0.01955690234899521, 0.15355655550956726, 0.04269038140773773, 0.029952816665172577, 0.07037873566150665, 0.022181103006005287, 0.020193306729197502, 0.17090263962745667, -0.0008546169847249985, 0.13909505307674408, 0.008186224848031998, -0.04685533419251442, -0.05492817610502243, -0.004771278705447912, -0.07576460391283035, 0.06035775691270828, 0.05392763018608093, -0.05940999090671539, -0.059477005153894424, -0.1314065009355545, 0.02757740207016468, 0.1591039001941681, 0.10289929062128067, -0.13187819719314575, -0.03020475246012211, 0.0693991482257843, 0.0225269366055727, -0.04749319329857826, 0.026629995554685593, -0.02790515311062336, -0.07501091808080673, 0.14926397800445557, 0.018853353336453438, 0.05834781378507614, 0.06438414007425308, 0.02890878915786743, -0.07046914845705032, -0.06939530372619629, 0.02126709371805191, 0.030984818935394287, -0.3025217056274414, 0.23893673717975616, -0.012847366742789745, -0.01568637415766716, -0.059326864778995514, 0.01222766563296318, 0.11124157905578613, 0.15374618768692017, 0.141686350107193, 0.03516387194395065, -0.08964737504720688, 0.05108082294464111, -0.029385332018136978, 0.03470449894666672, 0.013481189496815205, 0.031244775280356407, -0.018385160714387894, -0.05656033381819725, -0.049848150461912155, 0.04091273248195648, 0.35955002903938293, -0.0181563813239336, -0.14316405355930328, 0.00488900626078248, 0.05135079473257065, 0.002665963489562273, -0.039035093039274216, 0.003508813213557005, -0.08898892998695374, 0.1420067846775055, 0.04114149138331413, 0.04016347602009773, -0.10961548984050751, -0.06622058898210526, -0.00902164913713932, -0.040483493357896805, 0.08242624253034592, -0.04742693528532982, 0.008370144292712212, -0.06072365120053291, -0.17946326732635498, 0.12041398882865906, -0.1296832114458084, -0.02131393924355507, -0.09641941636800766, 0.0658242329955101, -0.07239050418138504, 0.028749510645866394, 0.02538529969751835, -0.0077821239829063416, -0.0545862652361393, -0.0257911067456007, 0.10931937396526337, 0.009990626946091652, -0.01783331111073494, 0.003358961781486869, -0.10273654013872147, -0.0778496265411377, 0.016271280124783516, -0.05587323009967804, 0.20155158638954163, 0.264487087726593, -0.036279160529375076, 0.14792096614837646, 0.14904798567295074, -0.06072483956813812, -0.3334738612174988, -0.11448083072900772, -0.15080402791500092, -0.0030955239199101925, -0.018887117505073547, -0.07749594748020172, 0.03075350448489189, -0.019915813580155373, -0.030651791021227837, 0.07945387065410614, 0.014758920297026634, -0.101986825466156, 0.18597351014614105, 0.021620944142341614, 0.36060699820518494, -0.10174008458852768, -0.02596992440521717, -0.1407444030046463, -0.17857396602630615, 0.1555066853761673, -0.09496951848268509, 0.07634812593460083, 0.006140653509646654, 0.020913582295179367, -0.029691921547055244, -0.01153622567653656, 0.09185738861560822, -0.0873638391494751, 0.01585582084953785, -0.17068108916282654, -0.005462192930281162, -0.017479825764894485, -0.04019203782081604, 0.06154277175664902, -0.09121756255626678, -0.003981107845902443, -0.06410937011241913, 0.013119845651090145, -0.10990936309099197, 0.11362852901220322, -0.012589279562234879, -0.09226778149604797, -0.05361732468008995, 0.004544415045529604, 0.06546613574028015, -0.018261205404996872, 0.08338992297649384, 0.003289071610197425, 0.12101184576749802, 0.07041703164577484, 0.05844821035861969, -0.1703179031610489, 0.004125709179788828, -0.02233687788248062, -0.10280906409025192, 0.00668461574241519, -0.1409745216369629, 0.05000491812825203, 0.026928342878818512, -0.04780220612883568, 0.08042247593402863, 0.07119623571634293, -0.03067406453192234, -0.04910023882985115, 0.15598222613334656, -0.13916365802288055, -0.05421053245663643, -0.02120635099709034, -0.03703876957297325, 0.014103097841143608, 0.00462064053863287, 0.1112029179930687, 0.00806653592735529, -0.014075756072998047, 0.0031989458948373795, 0.03903666511178017, -0.07704698294401169, 0.012997490353882313, 0.09840324521064758, -0.0302121639251709, -0.10053493827581406, 0.11057686805725098, -0.06797552108764648, -0.1951439529657364, 0.039958834648132324, 0.10669762641191483, -0.07382506132125854, -0.11606316268444061, 0.07807853817939758, 0.19164010882377625, -0.16105802357196808, -0.07638662308454514, -0.025371678173542023, -0.09320647269487381, 0.03683784604072571, 0.18594005703926086, 0.07071784883737564, 0.08612257987260818, 0.0602390319108963, -0.0067053912207484245, 0.0034209152217954397, 0.0029908171854913235, -0.07719044387340546, 0.041833724826574326, -0.0017751904670149088, -0.05166584625840187, -0.06684437394142151, 0.08251368999481201, -0.07780149579048157, 0.030334480106830597, -0.1400846391916275, -0.07888905704021454, -0.0992964655160904, -0.05410480499267578, -0.07843118906021118, -0.034284137189388275, 0.07254788279533386, -0.08509094268083572, -0.041477784514427185, -0.0671544149518013, -0.06678514927625656, -0.01803811453282833, 0.02949964441359043, 0.05962385609745979, -0.07725237309932709, -0.046928271651268005, 0.10912386327981949, -0.02630128525197506, 0.11504961550235748, 0.06957142800092697, -0.03969249501824379, 0.0704607218503952, -0.21025653183460236, -0.08165063709020615, 0.03980521112680435, -0.008736583404242992, 0.0928921103477478, -0.021153640002012253, -0.014889292418956757, 0.04019159451127052, -0.028428971767425537, 0.0432162880897522, 0.14723354578018188, -0.03106462024152279, 0.052221644669771194, 0.02587956003844738, -0.12266175448894501, 0.013020316138863564, -0.08674359321594238, 0.12964487075805664, -0.010719293728470802, 0.08751507848501205, -0.07782498002052307, 0.01337006501853466, -0.08663270622491837, 0.03571821376681328, -0.02049964666366577, -0.14637665450572968, -0.0018387750023975968, -0.03761456906795502, -0.0025712212081998587, 0.023035455495119095, 0.18070313334465027, -0.03630524501204491, -0.008936423808336258, 0.07375810295343399, 0.010824142955243587, 0.0029771258123219013, -0.025037292391061783, 0.21269561350345612, 0.001709701376967132, -0.06588558852672577, -0.09655490517616272, 0.0801360085606575, 0.04319365322589874, -0.011859051883220673, 0.06038876622915268, 0.023023543879389763, 0.07860353589057922, 0.13526102900505066, 0.0523679293692112, -0.029694339260458946, -0.004282251466065645, -0.2296971082687378, -0.017888901755213737, 0.08185743540525436, 0.03594531863927841, 0.09180285781621933, 0.08256598562002182, -0.10710500925779343, 0.05380251258611679, -0.08979471027851105, -0.02773798629641533, -0.12083067744970322, -0.04395734518766403, -0.09429043531417847, -0.10304245352745056, 0.029339520260691643, -0.06300444900989532, -0.027665603905916214, -0.06398086249828339, 0.040549878031015396, -0.055586740374565125, -0.008929513394832611, -0.0896850973367691, 0.0018335909117013216, 0.05077287182211876, 0.006419172044843435, -0.07875507324934006, -0.012185011990368366, -0.05709798261523247, -0.1256219446659088, 0.04495793581008911, -0.046658266335725784, 0.00486851017922163, -0.011063998565077782, 0.07859520614147186, -0.05201110616326332, -0.06583902984857559, -0.045135606080293655, -0.007067939732223749, -0.008694699965417385, 0.1269650161266327, 0.000644731568172574, 0.08899128437042236, 0.07619250565767288, 0.08464980125427246, -0.07301785796880722, -0.09356234967708588, -0.14672978222370148, 0.040147192776203156, -0.04030771926045418, 0.04833471402525902, -0.05625389143824577, 0.013657945208251476, -0.09093443304300308, 0.3029254674911499, 0.3131263852119446, -0.024368196725845337, 0.032140932977199554, -0.017036449164152145, 0.026934554800391197, -0.01527574472129345, 0.10911517590284348, 0.14582505822181702, 0.15664316713809967, -0.06075742840766907, -0.03567717969417572, -0.014862596057355404, -0.0029211752116680145, -0.07248326390981674, 0.06163349747657776, 0.037263546139001846, -0.05526388809084892, 0.010008626617491245, 0.03588821738958359, -0.10901358723640442, -0.017342885956168175, 0.042237788438797, -0.2267877459526062, -0.046092528849840164, -0.04195399209856987, 0.09461019933223724, 0.0740341767668724, 0.07978504151105881, 0.008572347462177277, 0.031598176807165146, 0.1407235562801361, -0.010245343670248985, -0.25555625557899475, -0.02603321522474289, 0.057902991771698, -0.1350136548280716, 0.17005322873592377, -0.02110816165804863, -0.028508298099040985, 0.10747753828763962, 0.04749493673443794, -0.06319935619831085, 0.07133786380290985, 0.01555169839411974, -0.029886770993471146, -0.04001736268401146, 0.025543980300426483, -0.025007689371705055, -0.09699390828609467, 0.04407969489693642, -0.03691478073596954, 0.06135052442550659, 0.10640411823987961, -0.057503629475831985, -0.045025359839200974, 0.11100033670663834, -0.13345268368721008, 0.09960591793060303, 0.0800287052989006, -0.009530874900519848, -0.028517015278339386, -0.04064209386706352, -0.0022162809036672115, 0.042081113904714584, -0.11924594640731812, -0.0655728206038475, -0.03067607432603836, -0.02045285515487194, -0.002362701576203108, -0.029108932241797447, -0.16161610186100006, -0.06919350475072861, -0.1366099715232849, -0.026403233408927917, -0.13976037502288818, 0.08264465630054474, 0.09535011649131775, 0.03793995454907417, 0.00045943690929561853, -0.04777298867702484, 0.04875069484114647, 0.07675064355134964, -0.0549234114587307, -0.12778374552726746 ]
null
null
transformers
# Bella Swan DialoGPT model
{"tags": ["conversational"]}
text-generation
Batsy24/DialoGPT-medium-Twilight_BellaBot
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Bella Swan DialoGPT model
[ "# Bella Swan DialoGPT model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Bella Swan DialoGPT model" ]
[ 51, 8 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Bella Swan DialoGPT model" ]
[ -0.025389065966010094, 0.06589903682470322, -0.006147907115519047, 0.03388277441263199, 0.10721607506275177, 0.008449924178421497, 0.10516903549432755, 0.11652238667011261, -0.08406049758195877, -0.024941053241491318, 0.11837157607078552, 0.1655428111553192, 0.003901456017047167, 0.10568921267986298, -0.07705675065517426, -0.3245309889316559, 0.0461922287940979, 0.032313816249370575, -0.014264595694839954, 0.11377853155136108, 0.08530458807945251, -0.04574406519532204, 0.06860599666833878, 0.034847985953092575, -0.11407794803380966, -0.006344295106828213, -0.016112638637423515, -0.10233347117900848, 0.1311979591846466, 0.06377444416284561, 0.052837081253528595, 0.04869542643427849, -0.03625424578785896, -0.1449337601661682, 0.03734367713332176, -0.04363255575299263, -0.03365283086895943, 0.0386991910636425, 0.011518614366650581, -0.05504186078906059, 0.14664380252361298, 0.12002415210008621, 0.040343012660741806, 0.04062732309103012, -0.12786193192005157, 0.03739842399954796, -0.006152470130473375, 0.08956967294216156, 0.042793773114681244, 0.09634019434452057, -0.03555436432361603, 0.08185986429452896, -0.10068012773990631, 0.08107807487249374, 0.023031631484627724, -0.32302677631378174, -0.030591165646910667, 0.0864696353673935, 0.07529198378324509, 0.06706931442022324, -0.03514367341995239, 0.1150617003440857, -0.0008411596645601094, 0.003090894315391779, -0.06997747719287872, -0.06689751893281937, -0.0951150432229042, -0.013013034127652645, -0.10321012884378433, 0.002820097142830491, 0.24743854999542236, -0.031189536675810814, 0.05377522110939026, -0.07813321799039841, -0.0834890604019165, -0.0027562829200178385, -0.05960403382778168, -0.07614900171756744, -0.07858691364526749, 0.08279459923505783, -0.003989171702414751, -0.09557808935642242, -0.10963523387908936, -0.012942611239850521, -0.1452367901802063, 0.11279714852571487, 0.02185272052884102, 0.04646263271570206, -0.22681556642055511, 0.07580913603305817, -0.06947340816259384, -0.06929273158311844, 0.030170319601893425, -0.09144094586372375, 0.052527427673339844, 0.004518403206020594, -0.03897140920162201, -0.05862325429916382, 0.07846508920192719, 0.1375359445810318, -0.008344635367393494, 0.017022136598825455, -0.009378669783473015, 0.043802715837955475, 0.039344023913145065, 0.009389002807438374, -0.028352083638310432, -0.13544529676437378, 0.046305347234010696, -0.09864822030067444, 0.02477942407131195, -0.06674094498157501, -0.15059176087379456, -0.0711045116186142, 0.007659023627638817, 0.030734697356820107, 0.04315362125635147, 0.09083428978919983, -0.031790271401405334, -0.020099682733416557, 0.04809226840734482, -0.010469304397702217, -0.001993325538933277, -0.00012162375060142949, 0.015399530529975891, 0.11628438532352448, 0.05764651671051979, 0.0660846158862114, -0.10245934128761292, 0.055089451372623444, -0.06223457679152489, -0.061734870076179504, 0.03825262933969498, -0.004202074836939573, -0.011408510617911816, -0.06762371212244034, 0.017574403434991837, -0.12829436361789703, -0.16682225465774536, 0.022982120513916016, 0.00024497698177583516, -0.07271981984376907, -0.14875344932079315, -0.09581311792135239, -0.03788305073976517, 0.07423944026231766, -0.035996127873659134, -0.00029401268693618476, -0.04460710287094116, 0.0675005316734314, -0.01050888653844595, 0.09260590374469757, -0.12327004969120026, 0.07322581857442856, -0.07713841646909714, -0.03866879642009735, -0.08381640166044235, 0.15947239100933075, 0.026018891483545303, 0.07129164040088654, -0.02130594290792942, -0.031968843191862106, -0.1434788852930069, 0.060788609087467194, -0.017910251393914223, 0.2442171573638916, -0.06022217497229576, -0.11123679578304291, 0.27757197618484497, -0.09797722101211548, -0.10483323037624359, 0.12844623625278473, 0.004697412718087435, 0.07863631844520569, 0.11489114910364151, 0.224916011095047, 0.02959383837878704, 0.049887143075466156, 0.10319407284259796, 0.07377933710813522, -0.08992654830217361, 0.0026683977339416742, -0.0016216777730733156, -0.037412118166685104, -0.032214146107435226, 0.032956186681985855, 0.0617569200694561, 0.040735118091106415, -0.03706596419215202, -0.04876989498734474, -0.027688510715961456, -0.023974338546395302, 0.09825777262449265, -0.03437982127070427, 0.11292793601751328, -0.04149807244539261, -0.059315986931324005, -0.05439268425107002, 0.019477833062410355, -0.042817290872335434, 0.049813784658908844, -0.06092900037765503, 0.10891525447368622, 0.04687533527612686, 0.06659270077943802, -0.13871844112873077, 0.029800428077578545, -0.05951765552163124, 0.198774516582489, 0.0681702047586441, 0.095867820084095, 0.04942698031663895, -0.028510121628642082, -0.026097729802131653, 0.02888035960495472, 0.1278180032968521, -0.022016601637005806, -0.06846918165683746, -0.11946059763431549, 0.08746769279241562, -0.054850365966558456, 0.14179883897304535, -0.13309317827224731, 0.011693411506712437, 0.02381625957787037, 0.08313753455877304, -0.01577620953321457, 0.02058599703013897, 0.009042240679264069, -0.023093178868293762, -0.06650735437870026, 0.04111333563923836, 0.091507188975811, -0.004870555829256773, -0.10872150957584381, 0.20534494519233704, -0.12870538234710693, 0.08259054273366928, 0.16570276021957397, -0.2599257230758667, -0.02713089808821678, -0.1036466434597969, -0.01168384775519371, 0.010888690128922462, 0.06475457549095154, -0.023897847160696983, 0.23206108808517456, -0.022353097796440125, 0.15459735691547394, -0.04334433749318123, 0.001310977153480053, -0.026090148836374283, -0.06326157599687576, -0.0020111831836402416, 0.09086970239877701, 0.07929811626672745, -0.08790349215269089, 0.12452270835638046, 0.04889308661222458, 0.0327293835580349, 0.2157510221004486, 0.05411545932292938, 0.022958973422646523, 0.0878443717956543, -0.005396803375333548, -0.0334673672914505, -0.08970530331134796, -0.34167996048927307, -0.03391510993242264, 0.06997638940811157, 0.01985827647149563, 0.09603995084762573, -0.11982278525829315, -0.05496053397655487, 0.013356149196624756, 0.00273873470723629, 0.03147462010383606, 0.0891784057021141, 0.008123393170535564, 0.12729987502098083, -0.015662875026464462, -0.07124373316764832, 0.061347559094429016, 0.01682111993432045, -0.08638562262058258, 0.18939654529094696, -0.11959648132324219, -0.2763867676258087, -0.1215713694691658, -0.13346120715141296, -0.0619410015642643, 0.0427611768245697, 0.10627467930316925, -0.10987254977226257, -0.023805376142263412, -0.018239453434944153, 0.0909908339381218, -0.1067463606595993, 0.012941261753439903, -0.007368482183665037, 0.006793472450226545, -0.1194513738155365, -0.09220074862241745, -0.0406465157866478, -0.05129920318722725, -0.055240198969841, 0.12715816497802734, -0.11625633388757706, 0.04969264566898346, 0.2504698634147644, 0.017307154834270477, 0.03689519688487053, -0.015656275674700737, 0.182870551943779, -0.10024343430995941, 0.027236003428697586, 0.18058501183986664, -0.02668360248208046, 0.08211228996515274, 0.16591067612171173, 0.002343012485653162, -0.08355198055505753, 0.02684856578707695, -0.026542875915765762, -0.072026327252388, -0.1907762736082077, -0.16009481251239777, -0.10625126957893372, 0.08172115683555603, -0.013571060262620449, 0.052751876413822174, 0.14535753428936005, 0.09297772496938705, -0.025951016694307327, -0.01467832736670971, 0.10930639505386353, 0.06970763951539993, 0.289287269115448, -0.05720747262239456, 0.16247527301311493, -0.007125393021851778, -0.14657624065876007, 0.059379253536462784, 0.06920959055423737, 0.07282446324825287, 0.05796493962407112, 0.05529087036848068, 0.03264414891600609, 0.04238249734044075, 0.1398632824420929, 0.05272357165813446, 0.027632467448711395, -0.05119239538908005, -0.026740126311779022, -0.03977183997631073, -0.02649279311299324, 0.04003647342324257, 0.049913376569747925, -0.11483070999383926, -0.028270430862903595, 0.061823420226573944, 0.04589938744902611, 0.05955122411251068, 0.09904484450817108, -0.1555025279521942, -0.017110735177993774, 0.062444671988487244, -0.05277560278773308, -0.0974382534623146, 0.06770209968090057, -0.0061638569459319115, -0.1327943354845047, 0.05622691661119461, -0.006413368973881006, 0.10239361971616745, -0.0650666207075119, 0.06511276960372925, -0.13841181993484497, -0.0747324600815773, -0.004031600430607796, 0.09772168844938278, -0.21394868195056915, 0.1880468875169754, -0.021763604134321213, -0.029009968042373657, -0.07522691786289215, 0.007864478044211864, 0.008087487891316414, 0.16510963439941406, 0.11767936497926712, -0.02192135527729988, 0.10483867675065994, 0.03182986378669739, -0.043108150362968445, 0.024590758606791496, 0.0899585708975792, -0.030731147155165672, -0.031019821763038635, -0.05463750660419464, 0.01664835773408413, -0.020806429907679558, 0.0031599272042512894, -0.008846933022141457, -0.20859545469284058, 0.08134294301271439, 0.029444606974720955, 0.05521779879927635, 0.04017430543899536, -0.034383781254291534, -0.003909159917384386, 0.24318376183509827, -0.023441849276423454, -0.10386136919260025, -0.08917971700429916, -0.03701053932309151, 0.046930283308029175, -0.05583525449037552, 0.0644305869936943, -0.06918537616729736, 0.01869632676243782, -0.11007104814052582, -0.14491203427314758, 0.08361008018255234, -0.08301916718482971, -0.0852026715874672, -0.037974126636981964, 0.178330197930336, -0.0067274379543960094, 0.022274570539593697, 0.07380052655935287, 0.017760569229722023, -0.11330042779445648, -0.07955976575613022, -0.029663702473044395, 0.02286211960017681, -0.005995390471071005, 0.012953314930200577, -0.016290830448269844, -0.08125050365924835, -0.048340290784835815, -0.02858663909137249, 0.3209030330181122, 0.15044493973255157, -0.042865827679634094, 0.18050923943519592, 0.1844143569469452, -0.055908720940351486, -0.31364068388938904, -0.10120825469493866, -0.07144109904766083, -0.044364940375089645, -0.0503012053668499, -0.1721334606409073, 0.11414453387260437, -0.014998081140220165, -0.013387751765549183, 0.11129332333803177, -0.3187254071235657, -0.101526640355587, 0.1392742544412613, -0.011615333147346973, 0.40418583154678345, -0.11436117440462112, -0.09534631669521332, -0.05477585271000862, -0.13510464131832123, 0.11727625876665115, 0.054961737245321274, 0.10598209500312805, -0.02217843197286129, 0.18946875631809235, 0.04125320538878441, 0.013482038863003254, 0.0777849331498146, 0.005918958690017462, -0.07867106050252914, -0.11226590722799301, -0.022333413362503052, 0.009165942668914795, 0.027110569179058075, 0.03976622223854065, -0.08409470319747925, 0.004796857014298439, -0.15082542598247528, -0.04092107713222504, -0.08210454881191254, 0.05260022357106209, 0.015950119122862816, -0.0705474391579628, -0.015273739583790302, -0.07212495058774948, 0.01658794656395912, 0.01477113924920559, 0.1341254860162735, -0.08850327879190445, 0.11307592689990997, 0.059167034924030304, 0.1619684100151062, -0.15138165652751923, -0.007216210942715406, -0.06438569724559784, -0.05706974118947983, 0.04904656112194061, -0.0973232313990593, 0.015547400340437889, 0.10174483060836792, -0.008627023547887802, 0.06415325403213501, 0.0914681926369667, -0.000673658330924809, 0.02569173276424408, 0.126495361328125, -0.27484244108200073, -0.042331475764513016, -0.054884396493434906, -0.01241379976272583, 0.06826438754796982, 0.10184814780950546, 0.1962219923734665, -0.010642686858773232, -0.0407380647957325, 0.005473911762237549, 0.05367032065987587, -0.06277380883693695, 0.09093094617128372, 0.004984538070857525, 0.0069551244378089905, -0.13937047123908997, 0.07622833549976349, 0.00798481423407793, -0.09502885490655899, 0.04759705811738968, 0.15183758735656738, -0.09853215515613556, -0.12231646478176117, -0.07464558631181717, 0.061549730598926544, -0.14247579872608185, 0.013579935766756535, -0.004421286750584841, -0.1421118825674057, 0.04058917239308357, 0.09661466628313065, 0.011636531911790371, 0.048345524817705154, -0.13173915445804596, -0.03172750771045685, -0.05645167455077171, 0.011180035769939423, 0.04670855402946472, -0.018330274149775505, -0.06411764770746231, 0.06847673654556274, -0.03691031411290169, 0.06203480064868927, -0.08166756480932236, -0.10481800138950348, -0.12428821623325348, 0.035452842712402344, -0.09211541712284088, -0.06704876571893692, -0.0939447209239006, -0.04345037415623665, -0.03680574521422386, 0.0027391258627176285, -0.031479839235544205, -0.04426423832774162, -0.0875505730509758, 0.019434168934822083, -0.04997763782739639, -0.003548929700627923, -0.09736558794975281, 0.03835112601518631, 0.04639541357755661, -0.01621689461171627, 0.14300447702407837, 0.12648795545101166, -0.09841831773519516, 0.04675845429301262, -0.1291917860507965, -0.06462981551885605, 0.09794074296951294, -0.015353478491306305, 0.04739077761769295, 0.09758281707763672, -0.008005290292203426, 0.041051313281059265, 0.02587500400841236, 0.06387840956449509, 0.08462008088827133, -0.08288750052452087, 0.03760670870542526, -0.059380702674388885, -0.16142277419567108, -0.035159073770046234, -0.029565555974841118, 0.08911006897687912, 0.016984406858682632, 0.10918858647346497, -0.06065737456083298, 0.058832328766584396, -0.0588141568005085, 0.031948551535606384, 0.025682484731078148, -0.12094210088253021, -0.02882358431816101, -0.08774133771657944, 0.02245458960533142, -0.01272760983556509, 0.16819503903388977, 0.019415391609072685, 0.005013569723814726, 0.005682297982275486, 0.08005359768867493, 0.057619206607341766, 0.019282810389995575, 0.21701152622699738, 0.11299610882997513, -0.04711237549781799, -0.10368618369102478, 0.10516157746315002, 0.041846830397844315, 0.0479624904692173, 0.08908703178167343, -0.04096946865320206, -0.050881482660770416, 0.10500767827033997, 0.014805952087044716, 0.06770789623260498, -0.12948544323444366, -0.15588437020778656, -0.030277738347649574, 0.050794631242752075, -0.04199718311429024, 0.11171621084213257, 0.15953192114830017, -0.039014268666505814, 0.01625289022922516, -0.010474232025444508, -0.08960941433906555, -0.18869173526763916, -0.24835501611232758, -0.06450435519218445, -0.12851904332637787, 0.012844469398260117, -0.14899973571300507, 0.02770104818046093, 0.026630869135260582, 0.07430638372898102, -0.08727205544710159, 0.05002940818667412, 0.0143054760992527, -0.0856974795460701, 0.08371441066265106, -0.0016621625982224941, 0.08107808232307434, 0.002117638010531664, 0.0002917940728366375, -0.05353029444813728, 0.05927519500255585, -0.004531259648501873, 0.05218430235981941, -0.07340346276760101, -0.001622658921405673, -0.10436279326677322, -0.08239057660102844, -0.04547896608710289, 0.0419357530772686, 0.008150637149810791, 0.14276769757270813, 0.0201713964343071, -0.0604848712682724, 0.020666811615228653, 0.21135158836841583, -0.04323948919773102, -0.03365855664014816, -0.06612960994243622, 0.15606537461280823, 0.004625630099326372, 0.14557158946990967, -0.018575703725218773, -0.022042933851480484, -0.0862012729048729, 0.31762930750846863, 0.3120773434638977, -0.10571349412202835, 0.020923510193824768, 0.03978444263339043, 0.04139294847846031, 0.1031418889760971, 0.08366435766220093, 0.10363438725471497, 0.2621310353279114, -0.05052531138062477, -0.003541277488693595, -0.05078963562846184, -0.04101544991135597, -0.06297627091407776, 0.05327669903635979, 0.05461642146110535, -0.05581776425242424, -0.01067687850445509, 0.1236630454659462, -0.24462123215198517, 0.09057597815990448, -0.10582214593887329, -0.21619294583797455, -0.0673280656337738, -0.026891134679317474, 0.0403696745634079, 0.02900112234055996, 0.0841117575764656, 0.02180313691496849, -0.05353347212076187, 0.044401925057172775, 0.03871655836701393, -0.1964821070432663, -0.022677384316921234, 0.0457434244453907, -0.08537685871124268, -0.020778406411409378, -0.02310650609433651, 0.005610917694866657, 0.06106838211417198, 0.0804382860660553, 0.005961920600384474, 0.03300495073199272, 0.003506291890516877, 0.016231289133429527, 0.012193947099149227, 0.07409563660621643, 0.016525808721780777, -0.09162785112857819, 0.10700351744890213, -0.13442838191986084, 0.04645565152168274, -0.026941383257508278, -0.007459822576493025, -0.012234621681272984, 0.06710730493068695, -0.08161501586437225, 0.07266473770141602, 0.09634639322757721, -0.010638889856636524, -0.00876978226006031, -0.03257124125957489, -0.021992750465869904, -0.018538841977715492, -0.05477755516767502, -0.07920736074447632, -0.20110954344272614, -0.11429610848426819, 0.062420833855867386, 0.03415684029459953, -0.1897507756948471, -0.004712516441941261, -0.14448580145835876, 0.06274029612541199, -0.13085289299488068, 0.10486558079719543, 0.13896852731704712, -0.0005602792371064425, 0.004406155552715063, -0.02042640745639801, 0.0592653863132, 0.11415937542915344, -0.13876239955425262, -0.05580129474401474 ]
null
null
transformers
# Twilight Edward DialoGPT Model
{"tags": ["conversational"]}
text-generation
Batsy24/DialoGPT-small-Twilight_EdBot
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Twilight Edward DialoGPT Model
[ "# Twilight Edward DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Twilight Edward DialoGPT Model" ]
[ 51, 8 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Twilight Edward DialoGPT Model" ]
[ -0.006624189671128988, 0.12125544995069504, -0.00566538330167532, 0.036583565175533295, 0.044069305062294006, 0.02130509540438652, 0.12659884989261627, 0.11493495851755142, -0.0440240241587162, -0.07358965277671814, 0.1426781266927719, 0.14429397881031036, -0.023236313834786415, 0.03490747883915901, -0.07678154855966568, -0.28877535462379456, 0.05892388895153999, 0.024981481954455376, 0.053669970482587814, 0.10339657962322235, 0.06165643408894539, -0.04995981231331825, 0.07106821984052658, -0.022848233580589294, -0.12646158039569855, -0.007736967410892248, -0.007646511308848858, -0.10756218433380127, 0.15558591485023499, 0.07261241972446442, 0.01323325838893652, 0.03185340017080307, -0.02938953973352909, -0.13767477869987488, 0.0359971784055233, -0.03233776241540909, -0.05045963078737259, 0.049733031541109085, 0.017011139541864395, -0.02166660502552986, 0.14655624330043793, 0.1488857865333557, 0.050338856875896454, 0.01732878014445305, -0.1350075751543045, -0.01637306995689869, 0.041978415101766586, 0.11427215486764908, 0.04414058104157448, 0.1257086992263794, -0.031616900116205215, 0.0907638669013977, -0.07861481606960297, 0.08126275986433029, 0.08548366278409958, -0.26452115178108215, -0.010550540871918201, 0.12770570814609528, 0.07683467864990234, 0.03810804709792137, -0.06153441220521927, 0.08116088062524796, 0.028035685420036316, 0.009330879896879196, 0.003967526368796825, -0.08684855699539185, -0.09041096270084381, -0.010286668315529823, -0.10281992703676224, -0.012476399540901184, 0.2175588756799698, -0.0605454221367836, 0.06252562254667282, -0.05383208766579628, -0.09292495250701904, 0.061420030891895294, -0.02554364688694477, -0.04208621382713318, -0.0807592123746872, 0.07553701847791672, 0.021342499181628227, -0.10178067535161972, -0.11799467355012894, -0.024178627878427505, -0.17592480778694153, 0.20723479986190796, 0.050334688276052475, 0.03766670450568199, -0.1872485727071762, 0.11523517221212387, -0.06965625286102295, -0.07685232162475586, 0.016342511400580406, -0.09317928552627563, 0.017316550016403198, 0.0027922382578253746, -0.04576420783996582, 0.026794733479619026, 0.05364720895886421, 0.2062181979417801, -0.055128466337919235, 0.011762064881622791, 0.024197950959205627, 0.0507226400077343, 0.04293837770819664, 0.0391414649784565, 0.031637776643037796, -0.09465958178043365, 0.052628133445978165, -0.09894813597202301, 0.027657916769385338, -0.050361257046461105, -0.18295599520206451, -0.07944314926862717, 0.037560321390628815, 0.00012015823449473828, 0.07036592066287994, 0.09444411844015121, -0.00042694498552009463, -0.047928813844919205, 0.040562260895967484, 0.0015905877808108926, -0.02527204528450966, 0.008886555209755898, 0.004222680814564228, 0.09997812658548355, 0.04393406957387924, 0.047491203993558884, -0.11368054896593094, 0.0987575352191925, -0.044191181659698486, -0.04964376613497734, 0.027196288108825684, -0.01933877542614937, 0.013203475624322891, -0.0010856406297534704, 0.025585798546671867, -0.12582767009735107, -0.17874431610107422, 0.03263426572084427, 0.027257075533270836, -0.029525037854909897, -0.15227185189723969, -0.12382014095783234, -0.02942853979766369, 0.059596385806798935, -0.059992242604494095, -0.012443073093891144, -0.05047890543937683, 0.06278569996356964, -0.01434273086488247, 0.09011541306972504, -0.1627804934978485, 0.06861719489097595, -0.1018325686454773, -0.050112172961235046, -0.06048281863331795, 0.0959499403834343, 0.05517207086086273, 0.05996977537870407, 0.010908985510468483, -0.025138070806860924, -0.05735065042972565, 0.07124994695186615, -0.05519101768732071, 0.2609306871891022, -0.08675133436918259, -0.11218059062957764, 0.2344650775194168, -0.08710278570652008, -0.15448004007339478, 0.1712617725133896, -0.005596430506557226, 0.02739744260907173, 0.11567386984825134, 0.20608769357204437, -0.03404899314045906, 0.0338299497961998, 0.09216535836458206, 0.1009795069694519, -0.09937626123428345, 0.014128941111266613, 0.03650326281785965, -0.002057020552456379, -0.05333636701107025, 0.020011993125081062, 0.05680161342024803, 0.0004373646806925535, -0.033082928508520126, -0.022414835169911385, -0.00028073787689208984, 0.013827024027705193, 0.1266622692346573, -0.020208535715937614, 0.10230006277561188, -0.06727239489555359, -0.054585348814725876, 0.010466346517205238, 0.03850012645125389, -0.034652549773454666, 0.079189732670784, -0.04434363916516304, 0.05927794426679611, -0.013829532079398632, 0.07002121210098267, -0.11474917829036713, 0.008698034100234509, -0.046746641397476196, 0.15543420612812042, 0.08464553207159042, 0.0904427245259285, 0.04595706984400749, -0.014376326464116573, -0.06695520132780075, 0.007386732380837202, 0.1276463121175766, -0.03151125833392143, -0.07023882120847702, -0.1584794819355011, 0.09408587217330933, -0.05610129237174988, 0.06408649682998657, -0.12372638285160065, 0.006819532252848148, -0.04711918532848358, 0.08587392419576645, -0.050362031906843185, 0.03703062981367111, -0.014796147122979164, -0.04972004517912865, -0.0867491364479065, 0.03209548443555832, 0.0810367539525032, -0.016209030523896217, -0.15381132066249847, 0.21596914529800415, -0.13538014888763428, 0.16341933608055115, 0.19833196699619293, -0.3137053847312927, 0.021459748968482018, -0.10025272518396378, -0.014017486944794655, -0.0063939932733774185, 0.057264238595962524, -0.04823431745171547, 0.3130400776863098, -0.02932608686387539, 0.15497104823589325, -0.04897962883114815, -0.054135315120220184, -0.05618510767817497, -0.036815643310546875, -0.012024345807731152, 0.11485288292169571, 0.05681821331381798, -0.08524646610021591, 0.14672912657260895, 0.0934990793466568, 0.04586222767829895, 0.1525266319513321, 0.0589357428252697, 0.014697614125907421, 0.08220680058002472, -0.07503066211938858, -0.04522746056318283, -0.07070675492286682, -0.2953203618526459, -0.01265754271298647, 0.07283105701208115, 0.03913659229874611, 0.10503149777650833, -0.10730690509080887, -0.06071494147181511, 0.01826016791164875, -0.0025148887652903795, -0.01897679828107357, 0.08148012310266495, 0.005063597112894058, 0.10889006406068802, 0.0018327084835618734, -0.0824676901102066, 0.07857553660869598, 0.0032257025595754385, -0.0924459844827652, 0.18125353753566742, -0.13871103525161743, -0.33725300431251526, -0.1405409276485443, -0.1423083394765854, -0.0680806115269661, 0.05824681743979454, 0.12146376073360443, -0.11271587014198303, -0.016724208369851112, -0.037511639297008514, 0.12356865406036377, -0.1696632355451584, 0.028073344379663467, -0.026923708617687225, 0.014995148405432701, -0.16665491461753845, -0.0780431479215622, -0.03767116367816925, -0.0483202300965786, -0.03373122587800026, 0.12108757346868515, -0.14994966983795166, 0.006645218003541231, 0.22238101065158844, 0.03537015616893768, 0.03170989081263542, -0.012291601859033108, 0.19098049402236938, -0.08746347576379776, -0.0023095877841115, 0.15351064503192902, -0.06774084270000458, 0.06169167533516884, 0.12463154643774033, 0.006055821198970079, -0.07629183679819107, 0.026693984866142273, -0.0441279299557209, -0.05777990072965622, -0.2129979431629181, -0.13972650468349457, -0.08626797795295715, 0.08578036725521088, 0.04445602744817734, 0.0454496368765831, 0.1898016333580017, 0.037084098905324936, -0.053869884461164474, 0.005594328511506319, 0.09508345276117325, 0.1276034712791443, 0.28352251648902893, -0.05590255185961723, 0.10113323479890823, 0.008876600302755833, -0.10778343677520752, 0.08623183518648148, 0.00685613788664341, 0.029795566573739052, 0.06433466821908951, 0.05322786420583725, -0.012205367907881737, -0.008099174126982689, 0.12667152285575867, 0.04729912802577019, 0.010669935494661331, -0.051259033381938934, -0.0352729856967926, -0.039650823920965195, -0.03177593648433685, 0.07843909412622452, 0.023485800251364708, -0.07857045531272888, -0.04977590963244438, -0.027792468667030334, 0.05280614271759987, 0.046064842492341995, 0.10877133905887604, -0.18865177035331726, -0.041434396058321, 0.07601863890886307, -0.07624460011720657, -0.08729301393032074, 0.07467660307884216, 0.03236749768257141, -0.11335653811693192, 0.06731809675693512, 0.015113560482859612, 0.1258748322725296, -0.09704915434122086, 0.08581298589706421, -0.15703651309013367, -0.09566860646009445, -0.013983403332531452, 0.08507637679576874, -0.27362143993377686, 0.21202820539474487, -0.020667340606451035, -0.019993843510746956, -0.06047322228550911, -0.028257349506020546, 0.010935571976006031, 0.16635073721408844, 0.11992869526147842, -0.00601562624797225, 0.11231495440006256, 0.08601893484592438, -0.08149097114801407, 0.03921167179942131, 0.07435328513383865, -0.07790980488061905, -0.02441910281777382, -0.04885663092136383, -0.004929603077471256, -0.007234298624098301, 0.009171606041491032, 0.02093036286532879, -0.156525656580925, 0.06866273283958435, 0.058306388556957245, 0.04884114861488342, 0.050314757972955704, -0.06109296530485153, -0.06439422070980072, 0.24793218076229095, -0.0490088164806366, -0.10033947974443436, -0.0841706171631813, -0.06766276061534882, 0.07196357101202011, -0.04426184296607971, 0.028036776930093765, -0.03936586156487465, 0.04467899352312088, -0.12027470767498016, -0.17439621686935425, 0.08578278869390488, -0.099934883415699, -0.08898867666721344, -0.055151138454675674, 0.20860625803470612, 0.013810070231556892, 0.06579285860061646, 0.05138177052140236, 0.017595358192920685, -0.1370525360107422, -0.0802624523639679, -0.011574986390769482, 0.07859490066766739, 0.003710169345140457, -0.01034227292984724, 0.015103921294212341, -0.04257438704371452, -0.026287013664841652, -0.026890907436609268, 0.2837879955768585, 0.1956951916217804, -0.008121106773614883, 0.16385893523693085, 0.20700103044509888, -0.0795648992061615, -0.2475978136062622, -0.09695757180452347, -0.11990734934806824, -0.0842011421918869, -0.10825284570455551, -0.14804115891456604, 0.0960412248969078, -0.02553846314549446, 0.017297884449362755, 0.05145867541432381, -0.30624616146087646, -0.11017366498708725, 0.1420547217130661, -0.008066571317613125, 0.4022083878517151, -0.07365497201681137, -0.07245149463415146, -0.08292775601148605, -0.1830003559589386, 0.1093447133898735, 0.04769715666770935, 0.10069648921489716, -0.05667737126350403, 0.16946843266487122, 0.039249103516340256, 0.010697562247514725, 0.06866757571697235, 0.01907951571047306, -0.056415095925331116, -0.10478055477142334, -0.0603380911052227, 0.0027790903113782406, -0.003472623648121953, 0.03752344846725464, -0.05593554303050041, 0.002703882986679673, -0.09628702700138092, -0.0508030503988266, -0.08269208669662476, 0.029209107160568237, -0.01459904108196497, -0.07600560784339905, 0.022739503532648087, -0.0473896749317646, 0.023327641189098358, 0.01765931211411953, 0.1746097356081009, -0.08250865340232849, 0.12249795347452164, -0.0011206913040950894, 0.1668550968170166, -0.13061606884002686, -0.026325436308979988, -0.03460613265633583, -0.05871223285794258, 0.06463110446929932, -0.11437835544347763, 0.0190040934830904, 0.1519152820110321, -0.004509340971708298, 0.058387406170368195, 0.11825007945299149, -0.005785848945379257, -0.006781621370464563, 0.08854992687702179, -0.29275721311569214, -0.057481978088617325, -0.06768665462732315, -0.06341920793056488, 0.08974367380142212, 0.10587792098522186, 0.17763882875442505, -0.048515018075704575, -0.054599471390247345, 0.027544839307665825, 0.030825847759842873, -0.04380066692829132, 0.0741351768374443, -0.03484675660729408, 0.016067417338490486, -0.15678299963474274, 0.12153603136539459, -0.027487823739647865, -0.12173140794038773, 0.02115703746676445, 0.21125559508800507, -0.08130750805139542, -0.11271488666534424, -0.039442747831344604, 0.049579232931137085, -0.17786364257335663, -0.007926350459456444, -0.02783484011888504, -0.10703469067811966, 0.08635583519935608, 0.12403517961502075, 0.045598093420267105, 0.04843826964497566, -0.1011163592338562, -0.012707272544503212, -0.026630723848938942, 0.017425063997507095, 0.03663521632552147, 0.012156368233263493, -0.040560171008110046, 0.0882011279463768, -0.043268512934446335, 0.07036508619785309, -0.09045795351266861, -0.09589902311563492, -0.1427994817495346, 0.049163561314344406, -0.13608016073703766, -0.07075663655996323, -0.0943363681435585, -0.03537214547395706, -0.024615999311208725, -0.017575126141309738, -0.017849694937467575, -0.02845073491334915, -0.09362442791461945, 0.00948353298008442, -0.026935365051031113, -0.012770012952387333, -0.08358024060726166, 0.011835061013698578, 0.05406269431114197, -0.027089577168226242, 0.1481064260005951, 0.13881371915340424, -0.11827818304300308, 0.0887162983417511, -0.16539016366004944, -0.07984576374292374, 0.09249492734670639, 0.017339812591671944, 0.0727008655667305, 0.08333764225244522, -0.037266671657562256, 0.04415521398186684, 0.024293994531035423, 0.0527171827852726, 0.07658849656581879, -0.0704009011387825, 0.01722869835793972, -0.010716593824326992, -0.17081046104431152, -0.019158290699124336, -0.060479771345853806, 0.07634121924638748, 0.013144150376319885, 0.09385914355516434, -0.04950527101755142, 0.09679042547941208, -0.04011944308876991, 0.0396980345249176, 0.04251660406589508, -0.1514364778995514, -0.0009060606244020164, -0.06234627589583397, 0.026861434802412987, -0.0029013853054493666, 0.1807362288236618, -0.02602975256741047, 0.03956887871026993, 0.027546387165784836, 0.02937988005578518, 0.010330328717827797, 0.0017447670688852668, 0.22675424814224243, 0.0925903171300888, -0.08001214265823364, -0.07600988447666168, 0.11185149103403091, 0.08761176466941833, 0.07496251910924911, 0.09810763597488403, -0.032822418957948685, 0.02104341983795166, 0.11322497576475143, 0.014381865039467812, 0.07801420986652374, -0.16894502937793732, -0.1835624724626541, 0.037210624665021896, 0.057844094932079315, -0.021328698843717575, 0.1424984633922577, 0.15099462866783142, -0.015197147615253925, 0.0129954032599926, -0.00944342091679573, -0.05362149327993393, -0.15701937675476074, -0.20980481803417206, -0.07980455458164215, -0.09023401141166687, -0.005655857268720865, -0.16352541744709015, 0.019600654020905495, -0.01157505251467228, 0.10502713918685913, -0.08459826558828354, 0.047024451196193695, 0.10943309962749481, -0.10431145876646042, 0.08423734456300735, -0.02435499243438244, 0.05483131855726242, -0.00842667929828167, 0.011859052814543247, -0.05021539703011513, 0.034528858959674835, -0.01128499023616314, 0.04238880053162575, -0.08188562840223312, 0.01616777293384075, -0.08320999890565872, -0.08820873498916626, -0.043879035860300064, 0.047765232622623444, 0.02088399976491928, 0.15796324610710144, 0.03929607570171356, -0.05777060613036156, 0.027817703783512115, 0.2551111876964569, -0.09012988209724426, -0.05562584102153778, -0.03380724415183067, 0.2557049095630646, 0.000603395514190197, 0.10359126329421997, -0.0017914047930389643, 0.00554799847304821, -0.10491297394037247, 0.29432258009910583, 0.2978254556655884, -0.07654397934675217, 0.0033572078682482243, 0.015638871118426323, 0.05044190213084221, 0.09395107626914978, 0.059708256274461746, 0.09768906980752945, 0.28437405824661255, -0.07081489264965057, -0.027755435556173325, -0.027793006971478462, 0.0008320836932398379, -0.004463103134185076, 0.03273793309926987, 0.06237998977303505, -0.07931672781705856, -0.0326288677752018, 0.09804162383079529, -0.2828176021575928, 0.060187745839357376, -0.1505984663963318, -0.19215747714042664, -0.04720764979720116, -0.006062173750251532, 0.05989118292927742, 0.02283354662358761, 0.09643968194723129, 0.031047506257891655, -0.07080078125, 0.013183250091969967, 0.009552151896059513, -0.21871264278888702, -0.04377860948443413, 0.08701130747795105, -0.07846857607364655, 0.0665503665804863, -0.015353037044405937, 0.06265059113502502, 0.04938345029950142, 0.07868316769599915, -0.017062759026885033, 0.016111653298139572, -0.002558989217504859, -0.056144796311855316, 0.01818377524614334, 0.09766657650470734, 0.01683799736201763, -0.03656788915395737, 0.11525905132293701, -0.12272832542657852, 0.034047164022922516, 0.01965334266424179, -0.05083596333861351, -0.013039728626608849, 0.027830561622977257, -0.07627315819263458, 0.07020467519760132, 0.09619928896427155, -0.01600380428135395, -0.034834057092666626, -0.03927744925022125, -0.015581590123474598, -0.021362418308854103, -0.03031681850552559, -0.06633119285106659, -0.16635724902153015, -0.08908355981111526, 0.039337072521448135, 0.0029101092368364334, -0.2020309418439865, -0.010715707205235958, -0.15042158961296082, 0.04568189010024071, -0.11948049813508987, 0.06699007004499435, 0.10565526783466339, 0.012083878740668297, 0.003681260859593749, -0.04238669201731682, 0.060132648795843124, 0.08749785274267197, -0.14930269122123718, -0.08946221321821213 ]
null
null
transformers
# DistilBERT with a second step of distillation ## Model description This model replicates the "DistilBERT (D)" model from Table 2 of the [DistilBERT paper](https://arxiv.org/pdf/1910.01108.pdf). In this approach, a DistilBERT student is fine-tuned on SQuAD v1.1, but with a BERT model (also fine-tuned on SQuAD v1.1) acting as a teacher for a second step of task-specific distillation. In this version, the following pre-trained models were used: * Student: `distilbert-base-uncased` * Teacher: `lewtun/bert-base-uncased-finetuned-squad-v1` ## Training data This model was trained on the SQuAD v1.1 dataset which can be obtained from the `datasets` library as follows: ```python from datasets import load_dataset squad = load_dataset('squad') ``` ## Training procedure ## Eval results | | Exact Match | F1 | |------------------|-------------|------| | DistilBERT paper | 79.1 | 86.9 | | Ours | 78.4 | 86.5 | The scores were calculated using the `squad` metric from `datasets`. ### BibTeX entry and citation info ```bibtex @misc{sanh2020distilbert, title={DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter}, author={Victor Sanh and Lysandre Debut and Julien Chaumond and Thomas Wolf}, year={2020}, eprint={1910.01108}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
{"language": ["en"], "license": "apache-2.0", "tags": ["question-answering"], "datasets": ["squad"], "metrics": ["squad"], "thumbnail": "https://github.com/karanchahal/distiller/blob/master/distiller.jpg"}
question-answering
BatuhanYilmaz/distilbert-base-uncased-finetuned-squad-d5716d28
[ "transformers", "pytorch", "distilbert", "fill-mask", "question-answering", "en", "dataset:squad", "arxiv:1910.01108", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "1910.01108" ]
[ "en" ]
TAGS #transformers #pytorch #distilbert #fill-mask #question-answering #en #dataset-squad #arxiv-1910.01108 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
DistilBERT with a second step of distillation ============================================= Model description ----------------- This model replicates the "DistilBERT (D)" model from Table 2 of the DistilBERT paper. In this approach, a DistilBERT student is fine-tuned on SQuAD v1.1, but with a BERT model (also fine-tuned on SQuAD v1.1) acting as a teacher for a second step of task-specific distillation. In this version, the following pre-trained models were used: * Student: 'distilbert-base-uncased' * Teacher: 'lewtun/bert-base-uncased-finetuned-squad-v1' Training data ------------- This model was trained on the SQuAD v1.1 dataset which can be obtained from the 'datasets' library as follows: Training procedure ------------------ Eval results ------------ Exact Match: DistilBERT paper, F1: 79.1 Exact Match: Ours, F1: 78.4 The scores were calculated using the 'squad' metric from 'datasets'. ### BibTeX entry and citation info
[ "### BibTeX entry and citation info" ]
[ "TAGS\n#transformers #pytorch #distilbert #fill-mask #question-answering #en #dataset-squad #arxiv-1910.01108 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### BibTeX entry and citation info" ]
[ 68, 11 ]
[ "passage: TAGS\n#transformers #pytorch #distilbert #fill-mask #question-answering #en #dataset-squad #arxiv-1910.01108 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### BibTeX entry and citation info" ]
[ -0.07598058879375458, 0.1573057472705841, -0.005492530297487974, 0.09660123288631439, 0.036158982664346695, 0.03898710384964943, 0.16792051494121552, 0.123894102871418, 0.05372770130634308, -0.020512646064162254, 0.1637255698442459, 0.20980204641819, 0.044715095311403275, 0.13357822597026825, -0.10912513732910156, -0.1332099586725235, 0.01575417071580887, 0.08974277973175049, 0.041118789464235306, 0.0906098335981369, 0.0895283967256546, -0.06568839401006699, 0.049830224364995956, -0.035770729184150696, -0.06624606996774673, 0.030915869399905205, 0.025108860805630684, -0.06645005941390991, 0.11519787460565567, 0.06876041740179062, 0.039744965732097626, 0.07009582221508026, 0.01552195567637682, -0.11354787647724152, 0.041836049407720566, -0.020993288606405258, -0.07611622661352158, 0.12164946645498276, -0.02541547454893589, -0.029316095635294914, -0.007006686180830002, 0.02482086978852749, -0.019850721582770348, 0.03598823770880699, -0.10522842407226562, -0.18955068290233612, -0.11401132494211197, 0.14090633392333984, 0.029600244015455246, 0.044157158583402634, 0.043979283422231674, 0.17662183940410614, -0.029501130804419518, 0.0790165439248085, 0.16624608635902405, -0.30488747358322144, -0.012509343214333057, -0.007375533226877451, 0.05662050098180771, -0.0036145455669611692, -0.040623877197504044, 0.022253340110182762, 0.08074326813220978, 0.01066061295568943, 0.012968827039003372, -0.09217073023319244, -0.12268223613500595, 0.007294336799532175, -0.02745659090578556, -0.05831404775381088, 0.23456831276416779, 0.01726916991174221, -0.03802212327718735, 0.027843791991472244, -0.06733831018209457, 0.09592815488576889, -0.02986934408545494, 0.035742681473493576, 0.005023249890655279, 0.012640629895031452, 0.06362170726060867, -0.05377885699272156, -0.11485523730516434, -0.05891509726643562, -0.1708635687828064, 0.05022606626152992, -0.005893401335924864, 0.093903549015522, -0.12086643278598785, 0.04676153510808945, 0.027124734595417976, -0.1404813826084137, -0.008428141474723816, -0.060233041644096375, 0.09133269637823105, 0.023619530722498894, 0.021781323477625847, 0.06609030812978745, 0.12238762527704239, 0.30641692876815796, 0.054231688380241394, -0.02777453139424324, -0.11451376229524612, 0.10257403552532196, 0.01726013608276844, 0.012073541060090065, -0.01381711382418871, -0.08295227587223053, 0.1290355920791626, -0.04961453750729561, 0.09568451344966888, -0.02172575518488884, -0.1341456174850464, -0.05903683975338936, -0.017462119460105896, 0.10056093335151672, 0.134485125541687, 0.001483079046010971, -0.04905678704380989, -0.000037482524930965155, 0.1980622261762619, -0.07012517005205154, 0.00268527097068727, -0.011172556318342686, 0.004766094032675028, -0.04402042180299759, 0.07347071915864944, 0.018451442942023277, -0.05280125141143799, 0.06695227324962616, -0.06987111270427704, -0.0457427054643631, 0.018234889954328537, -0.05655946582555771, 0.0924878939986229, -0.12164513766765594, 0.06863490492105484, -0.13889379799365997, -0.22315816581249237, 0.04665061831474304, 0.05739126354455948, -0.07055763900279999, -0.07116971164941788, 0.08463065326213837, 0.02098018489778042, 0.038073793053627014, -0.07952864468097687, 0.0536775067448616, -0.08769849687814713, 0.09367749840021133, -0.040873028337955475, 0.08242450654506683, -0.20098236203193665, 0.037076011300086975, -0.120749831199646, -0.0024240254424512386, 0.02006281353533268, -0.03876866400241852, -0.09033462405204773, 0.06839430332183838, -0.11090140789747238, -0.018250631168484688, -0.02734324149787426, -0.028064558282494545, 0.05600326508283615, 0.15687376260757446, -0.13678114116191864, -0.03614801540970802, 0.17390574514865875, -0.08176761865615845, -0.2455374300479889, 0.08961565792560577, -0.015313567593693733, 0.0581057146191597, 0.057160068303346634, 0.10478939116001129, 0.018259745091199875, -0.11812714487314224, 0.01037020143121481, 0.096162810921669, -0.0375915989279747, -0.1386786252260208, 0.05192284286022186, 0.001941001508384943, -0.055669158697128296, 0.0328538715839386, 0.07776409387588501, 0.02960296720266342, -0.014882064424455166, -0.08336834609508514, -0.04730004817247391, -0.04680272191762924, 0.0035429091658443213, 0.04078911989927292, 0.022471629083156586, -0.06453610211610794, 0.007993526756763458, -0.027543429285287857, 0.023333247750997543, 0.08066083490848541, 0.01739124208688736, -0.020117033272981644, 0.08219017833471298, -0.12332799285650253, 0.028681393712759018, -0.13239648938179016, 0.06647989898920059, -0.015241392888128757, 0.045569367706775665, 0.060538310557603836, 0.053464051336050034, 0.04234655946493149, -0.04036126658320427, -0.028760667890310287, -0.0165301114320755, 0.09603098034858704, 0.041008416563272476, -0.07031368464231491, -0.16778214275836945, 0.013983091339468956, -0.04877486452460289, -0.024997659027576447, -0.03908054158091545, 0.00879753939807415, -0.0630953460931778, 0.07644766569137573, -0.02855217270553112, 0.08857572078704834, -0.012968234717845917, -0.0342472679913044, -0.06726064532995224, 0.0013117357157170773, 0.08525111526250839, 0.033267680555582047, -0.03954076021909714, 0.1377546340227127, -0.0066131288185715675, 0.19409777224063873, 0.16859343647956848, -0.09788284450769424, 0.021330999210476875, -0.046956516802310944, -0.04962991923093796, -0.04370596632361412, 0.014451373368501663, 0.013027625158429146, -0.057072192430496216, 0.00922441016882658, 0.1194082647562027, -0.08786434680223465, -0.0344315767288208, 0.021152092143893242, -0.07159312069416046, -0.016700198873877525, 0.05881102755665779, 0.15749095380306244, -0.23339632153511047, 0.1843709796667099, 0.25689896941185, -0.002575851744040847, 0.11429335922002792, -0.013400835916399956, -0.05576260760426521, -0.01738697849214077, -0.07656088471412659, -0.00259336712770164, 0.12740637362003326, -0.06894809007644653, 0.06432077288627625, 0.12010280042886734, -0.03567036986351013, 0.01591876894235611, -0.10488379001617432, -0.041552942246198654, -0.02852548472583294, -0.0028753799851983786, -0.10679007321596146, 0.07027338445186615, 0.012646762654185295, 0.14763042330741882, 0.0029952540062367916, -0.0480528399348259, 0.07706198841333389, -0.008639211766421795, -0.08757899701595306, 0.15035443007946014, -0.10214044898748398, -0.2606252431869507, -0.12003425508737564, -0.14176365733146667, -0.01470060832798481, 0.02043294720351696, 0.11439816653728485, 0.007791904732584953, -0.02557510882616043, 0.02118358574807644, -0.07412370294332504, 0.03357456251978874, -0.010674807243049145, -0.040865641087293625, 0.05257368087768555, 0.017097435891628265, -0.1327321082353592, -0.04656641557812691, -0.009457890875637531, 0.02890584245324135, 0.0825524851679802, -0.06238695979118347, 0.08281825482845306, 0.031467437744140625, -0.0021678684279322624, 0.0030543156899511814, -0.023677295073866844, 0.1955767720937729, -0.03090279921889305, 0.07786083966493607, 0.20373372733592987, -0.048883069306612015, 0.061960168182849884, 0.19379793107509613, 0.03758732229471207, -0.04827011004090309, 0.01683248206973076, -0.046403974294662476, -0.067514568567276, -0.2800613045692444, -0.06281383335590363, -0.11343737691640854, 0.06764566153287888, 0.11280227452516556, 0.03445233404636383, 0.12869258224964142, 0.09613022953271866, -0.05683054402470589, 0.07808373123407364, -0.008075552061200142, 0.07541050761938095, 0.15477174520492554, -0.006035580765455961, 0.12640567123889923, -0.09881865233182907, -0.03334907069802284, 0.09578932076692581, 0.14592686295509338, 0.08555544167757034, 0.07326199114322662, 0.07701867818832397, 0.08458134531974792, 0.09899768978357315, 0.06330318003892899, 0.15781250596046448, 0.01193223800510168, -0.0049486951902508736, -0.04229277744889259, -0.03816258907318115, -0.0486331433057785, 0.05295339971780777, -0.011396337300539017, 0.026694532483816147, -0.0024907244369387627, -0.1353870928287506, 0.025112463161349297, 0.1650669425725937, 0.08799415826797485, -0.15474000573158264, -0.0481339767575264, 0.07801028341054916, 0.01832546293735504, -0.08649364858865738, 0.0058942134492099285, -0.07784128189086914, -0.10467185825109482, 0.05774786323308945, -0.024127621203660965, 0.14069713652133942, 0.03231331333518028, 0.013593016192317009, -0.1034935861825943, -0.10324440151453018, 0.0016467984532937407, 0.13420763611793518, -0.30352309346199036, 0.23544374108314514, 0.04068101570010185, -0.019022412598133087, -0.10025713592767715, 0.009422573260962963, 0.01269945502281189, 0.08161035925149918, 0.14286278188228607, -0.011960377916693687, 0.02678236924111843, -0.026696791872382164, -0.0659153014421463, 0.05143168568611145, -0.060124631971120834, -0.009399537928402424, -0.028365997597575188, -0.014904421754181385, -0.008568177931010723, -0.00633784756064415, 0.0632551908493042, -0.08220488578081131, -0.15786951780319214, 0.06520242989063263, 0.046475525945425034, -0.0780852809548378, -0.03772563487291336, -0.04697102680802345, 0.011915823444724083, 0.1716790497303009, -0.14137567579746246, -0.08094514161348343, -0.10242129862308502, 0.037833768874406815, 0.09935072064399719, -0.08293458819389343, 0.04583699256181717, -0.09134938567876816, -0.002978530013933778, -0.0855947807431221, -0.1773335486650467, 0.08554552495479584, -0.14393584430217743, -0.07556812465190887, -0.08381423354148865, 0.1389150470495224, -0.06936773657798767, 0.04676400497555733, 0.039566583931446075, 0.021992454305291176, -0.1182996928691864, -0.0683315321803093, 0.0463721938431263, -0.09767383337020874, 0.16914266347885132, -0.011908404529094696, -0.10724151879549026, -0.12874926626682281, 0.009533645585179329, -0.04431864246726036, 0.1633341908454895, 0.23088663816452026, -0.07846692949533463, 0.1671133190393448, 0.21278385818004608, -0.08110874146223068, -0.2624484598636627, -0.10774488002061844, -0.10511621087789536, -0.05114834010601044, -0.019119713455438614, -0.10691528022289276, 0.10880158096551895, 0.0492195226252079, -0.08516435325145721, 0.15761779248714447, -0.09699850529432297, -0.07700076699256897, 0.23597972095012665, -0.0034569944255053997, 0.26081448793411255, -0.13055458664894104, -0.08454474806785583, -0.115155428647995, -0.26521146297454834, 0.1464454084634781, -0.043647702783346176, 0.045543741434812546, -0.07282452285289764, -0.007047454826533794, -0.026763461530208588, -0.05571123957633972, 0.11028720438480377, -0.033518869429826736, 0.01663305051624775, -0.1029517725110054, -0.058387260884046555, 0.029113998636603355, -0.0037192469462752342, 0.0839613601565361, -0.11165060847997665, 0.05714774504303932, -0.09489301592111588, -0.013363388366997242, -0.06803157925605774, 0.10788637399673462, -0.031496383249759674, -0.0839824452996254, -0.03912048786878586, -0.001423945534043014, -0.03512369096279144, -0.014441940933465958, 0.1954280436038971, 0.012500791810452938, 0.13082881271839142, 0.15525057911872864, 0.07501787692308426, -0.1425313800573349, 0.016780911013484, -0.08652780205011368, -0.09014251828193665, 0.06285100430250168, -0.11223945021629333, 0.06738274544477463, 0.12713013589382172, 0.008716699667274952, 0.09430922567844391, 0.0448380708694458, -0.0041931550949811935, -0.0343417190015316, 0.08428405225276947, -0.181634321808815, 0.06567863374948502, 0.03054751269519329, 0.054110921919345856, -0.03315528482198715, 0.06135343760251999, 0.11467056721448898, 0.0038553315680474043, -0.0827297791838646, 0.03860313072800636, 0.06907474249601364, -0.04086323827505112, 0.11270099133253098, 0.1254708170890808, 0.030818235129117966, -0.11999185383319855, 0.07551521062850952, 0.03565375134348869, -0.06395699828863144, -0.007310437969863415, 0.045858483761548996, -0.06695912778377533, -0.11877606064081192, 0.025299007073044777, 0.026291677728295326, -0.08131857216358185, -0.059862542897462845, -0.05194849520921707, -0.08213925361633301, 0.07259400933980942, 0.1704917550086975, 0.06159800663590431, 0.01603744924068451, 0.004434096626937389, -0.07538814097642899, 0.006379649043083191, 0.09977811574935913, -0.06236063688993454, 0.03800191730260849, 0.0046927593648433685, -0.10426078736782074, -0.0063285124488174915, 0.13450005650520325, -0.04851243644952774, 0.023366065695881844, -0.14230144023895264, 0.03506085276603699, -0.21193371713161469, 0.047497306019067764, -0.10612994432449341, -0.019714057445526123, -0.037378210574388504, -0.10208362340927124, -0.05899282172322273, -0.011398916132748127, -0.08689121901988983, 0.00224303244613111, 0.01324793417006731, 0.07052820175886154, -0.15917225182056427, -0.061376359313726425, 0.10102581232786179, -0.006872606463730335, 0.09602886438369751, 0.08388033509254456, -0.030804188922047615, 0.030894851312041283, -0.16079308092594147, -0.0725315511226654, 0.043453339487314224, 0.054099127650260925, 0.06613799184560776, -0.04258345440030098, 0.013613534159958363, 0.09782524406909943, -0.03208395466208458, 0.01706071011722088, 0.053034357726573944, -0.11625343561172485, -0.037687841802835464, -0.04085947573184967, -0.12009090930223465, -0.016706397756934166, -0.04275361821055412, 0.08886581659317017, 0.020925337448716164, 0.10978597402572632, -0.02935590222477913, 0.048323675990104675, -0.14216338098049164, 0.013917878270149231, -0.05621621757745743, -0.14501184225082397, -0.08740530908107758, -0.042252011597156525, -0.00827660784125328, -0.01864021271467209, 0.24910302460193634, -0.009846254251897335, -0.01975037157535553, 0.03556936979293823, 0.04998452588915825, -0.01463017426431179, -0.011466861702501774, 0.21428357064723969, -0.0005723568028770387, -0.015821080654859543, -0.0962895080447197, 0.04496563971042633, -0.04037817567586899, 0.029985422268509865, 0.08072244375944138, 0.13590656220912933, 0.1274414360523224, 0.03467641770839691, 0.0782492384314537, -0.01666051335632801, 0.016065344214439392, -0.12429873645305634, 0.06056009605526924, 0.05908973142504692, 0.05625343322753906, 0.09674309194087982, 0.1399870067834854, -0.05602826178073883, 0.014261198230087757, -0.0626022070646286, -0.020355239510536194, -0.13937389850616455, -0.0944853350520134, -0.06211983039975166, -0.02038807049393654, -0.00848991796374321, -0.07030591368675232, -0.005116832442581654, 0.030845653265714645, 0.037179332226514816, -0.054615553468465805, -0.026345089077949524, 0.04437202587723732, -0.02708321064710617, 0.01588420942425728, 0.03554641827940941, -0.017301037907600403, -0.06303402036428452, 0.04051756486296654, -0.03494224697351456, 0.014093955978751183, -0.015437157824635506, 0.020409349352121353, 0.01918739639222622, 0.03423325717449188, -0.08294427394866943, -0.09858470410108566, -0.052931904792785645, 0.03730052709579468, 0.033921562135219574, 0.18754780292510986, 0.02413805201649666, 0.09018383175134659, 0.08338670432567596, 0.12065273523330688, -0.028161292895674706, -0.12867332994937897, -0.0882946252822876, 0.14846959710121155, -0.02870035171508789, 0.02791045419871807, 0.003684654366225004, -0.011845477856695652, 0.003721728688105941, 0.24406595528125763, 0.2953552007675171, -0.06473667919635773, 0.040155816823244095, 0.04005530849099159, 0.011880439706146717, 0.06505746394395828, 0.04634292051196098, 0.11075233668088913, 0.20924797654151917, -0.07124147564172745, -0.023800117895007133, -0.0592404268682003, 0.0017269187374040484, -0.04409648850560188, 0.10178860276937485, -0.008754352107644081, -0.05777301639318466, -0.009286348707973957, 0.07649700343608856, -0.041465312242507935, -0.06975255161523819, 0.005824678111821413, -0.13540500402450562, -0.09414549916982651, 0.004759438335895538, 0.03755516558885574, 0.02029437944293022, -0.007005095481872559, -0.046089716255664825, -0.007369011174887419, 0.07605813443660736, -0.025444338098168373, -0.15548503398895264, -0.07910747826099396, 0.09374100714921951, -0.027675818651914597, 0.13809116184711456, -0.0259417574852705, 0.10621322691440582, 0.09102782607078552, 0.0643969178199768, -0.11569580435752869, 0.04949166998267174, 0.04983556270599365, -0.0037574588786810637, 0.05281871184706688, -0.06643977016210556, 0.009475508704781532, -0.08315324038267136, 0.0947265699505806, -0.04915481060743332, 0.02615482546389103, 0.00270947627723217, -0.0954682007431984, -0.08310161530971527, 0.11611536145210266, -0.08283354341983795, 0.07130325585603714, 0.01808888278901577, -0.06776561588048935, -0.07539758831262589, -0.06513608247041702, 0.016134263947606087, 0.07504446804523468, -0.10731161385774612, -0.0533195398747921, -0.02846664749085903, 0.01944848708808422, -0.029851844534277916, 0.04876858368515968, -0.2152537852525711, -0.022024638950824738, -0.1031208261847496, -0.04534538462758064, -0.12557587027549744, 0.015130976215004921, 0.05614713951945305, 0.016964031383395195, -0.02051788568496704, -0.1243930384516716, 0.005045443307608366, 0.014323997311294079, -0.09389553219079971, -0.06321050226688385 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # dummy-model This model is a fine-tuned version of [camembert-base](https://huggingface.co/camembert-base) on an unknown dataset. It achieves the following results on the evaluation set: ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: None - training_precision: float32 ### Training results ### Framework versions - Transformers 4.15.0 - TensorFlow 2.7.0 - Datasets 1.17.0 - Tokenizers 0.10.3
{"license": "mit", "tags": ["generated_from_keras_callback"], "model-index": [{"name": "dummy-model", "results": []}]}
fill-mask
BatuhanYilmaz/dummy-model
[ "transformers", "tf", "camembert", "fill-mask", "generated_from_keras_callback", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #tf #camembert #fill-mask #generated_from_keras_callback #license-mit #autotrain_compatible #endpoints_compatible #region-us
# dummy-model This model is a fine-tuned version of camembert-base on an unknown dataset. It achieves the following results on the evaluation set: ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: None - training_precision: float32 ### Training results ### Framework versions - Transformers 4.15.0 - TensorFlow 2.7.0 - Datasets 1.17.0 - Tokenizers 0.10.3
[ "# dummy-model\n\nThis model is a fine-tuned version of camembert-base on an unknown dataset.\nIt achieves the following results on the evaluation set:", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- optimizer: None\n- training_precision: float32", "### Training results", "### Framework versions\n\n- Transformers 4.15.0\n- TensorFlow 2.7.0\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #tf #camembert #fill-mask #generated_from_keras_callback #license-mit #autotrain_compatible #endpoints_compatible #region-us \n", "# dummy-model\n\nThis model is a fine-tuned version of camembert-base on an unknown dataset.\nIt achieves the following results on the evaluation set:", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- optimizer: None\n- training_precision: float32", "### Training results", "### Framework versions\n\n- Transformers 4.15.0\n- TensorFlow 2.7.0\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ 53, 39, 6, 12, 8, 3, 33, 4, 31 ]
[ "passage: TAGS\n#transformers #tf #camembert #fill-mask #generated_from_keras_callback #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# dummy-model\n\nThis model is a fine-tuned version of camembert-base on an unknown dataset.\nIt achieves the following results on the evaluation set:## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- optimizer: None\n- training_precision: float32### Training results### Framework versions\n\n- Transformers 4.15.0\n- TensorFlow 2.7.0\n- Datasets 1.17.0\n- Tokenizers 0.10.3" ]
[ -0.046263109892606735, -0.013249280862510204, -0.0010372534161433578, 0.07224117964506149, 0.16367055475711823, 0.029854383319616318, 0.12861892580986023, 0.09136445820331573, -0.12264850735664368, 0.0014719413593411446, 0.08226197212934494, 0.12172068655490875, 0.014069318771362305, 0.09793684631586075, -0.03790612891316414, -0.24302507936954498, 0.025157473981380463, 0.0028308636974543333, -0.08009346574544907, 0.09441997855901718, 0.08686142414808273, -0.10544568300247192, 0.09119243174791336, 0.015890154987573624, -0.24552203714847565, 0.024596042931079865, 0.03637617081403732, -0.06676638126373291, 0.11250047385692596, 0.03458351269364357, 0.12410704791545868, 0.020638398826122284, 0.10992175340652466, -0.1105632558465004, 0.02048715017735958, 0.08763877302408218, 0.005260271951556206, 0.0647633895277977, -0.018024200573563576, -0.004626147914677858, 0.18375715613365173, -0.08492272347211838, 0.09815899282693863, 0.03861729055643082, -0.12043581157922745, -0.14226415753364563, -0.0776231437921524, -0.025512458756566048, 0.032986223697662354, 0.12459401041269302, -0.0024791026953607798, 0.2643175721168518, -0.0986390933394432, 0.07180636376142502, 0.1407320201396942, -0.2822648882865906, -0.07543836534023285, 0.10820271819829941, 0.0672932043671608, 0.014648210257291794, -0.08195782452821732, 0.03778344392776489, 0.06749662011861801, 0.054043181240558624, 0.09231134504079819, -0.041366055607795715, -0.16092775762081146, -0.014019342139363289, -0.1099776029586792, 0.03823245316743851, 0.143455371260643, 0.01121379155665636, -0.05907829478383064, -0.03564073145389557, -0.04286086559295654, -0.028236988931894302, -0.031067339703440666, -0.07274410873651505, 0.05466622859239578, -0.0025170939043164253, -0.09205175191164017, -0.058277834206819534, -0.0932929664850235, -0.08204862475395203, -0.1284961849451065, 0.1420450210571289, -0.0024227402172982693, 0.043567489832639694, -0.10473500937223434, 0.06572001427412033, -0.11134009063243866, -0.09954354912042618, -0.004380379803478718, -0.0073081450536847115, -0.0712149366736412, -0.09854297339916229, -0.09775847941637039, -0.18250317871570587, 0.04191124439239502, 0.09615366160869598, -0.06169627234339714, 0.06923321634531021, -0.06423540413379669, 0.039692506194114685, -0.014870009385049343, 0.12119456380605698, -0.05515298992395401, 0.029838476330041885, 0.009361662901937962, -0.014472399838268757, 0.005022069904953241, -0.013839724473655224, -0.10050498694181442, 0.014541144482791424, 0.022432420402765274, 0.01329079084098339, -0.058356791734695435, 0.08601800352334976, -0.04988056421279907, 0.003583529032766819, -0.009920173324644566, -0.07667884975671768, 0.022514548152685165, -0.032627176493406296, -0.07666970044374466, 0.027378397062420845, 0.10477186739444733, 0.01632779650390148, 0.004574764519929886, 0.0725921243429184, -0.08897511661052704, 0.020821886137127876, -0.11273381859064102, -0.10661349445581436, 0.005013900343328714, -0.08595354855060577, 0.02886885218322277, -0.08060264587402344, -0.1945846974849701, 0.0007916621398180723, 0.10129941254854202, -0.07366402447223663, 0.04821957275271416, -0.05665231868624687, -0.07591227442026138, 0.003753518220037222, 0.008985774591565132, 0.14890407025814056, -0.047725461423397064, 0.03750481829047203, 0.00146473350469023, 0.08958957344293594, -0.09498803317546844, 0.02537045255303383, -0.09059730917215347, 0.003247619839385152, -0.17471598088741302, 0.07505975663661957, -0.05120540410280228, 0.07327596098184586, -0.10418793559074402, -0.07874363660812378, -0.04869521036744118, 0.01724928617477417, 0.08313815295696259, 0.12370943278074265, -0.23923331499099731, -0.019118433818221092, 0.15213432908058167, -0.11310355365276337, -0.11137361824512482, 0.06640312075614929, -0.08999963104724884, 0.18516410887241364, 0.06978695839643478, 0.11452256888151169, -0.012979773804545403, -0.17564913630485535, 0.06608696281909943, 0.011381756514310837, -0.048381682485342026, 0.017602358013391495, -0.015517224557697773, -0.006511122453957796, -0.04923788085579872, 0.02003004215657711, -0.002897280501201749, 0.014528325758874416, -0.10693833231925964, -0.07398667931556702, -0.046797655522823334, -0.08875440806150436, 0.06322532147169113, 0.022582178935408592, 0.09426435828208923, -0.07432658970355988, -0.10669029504060745, 0.12890909612178802, 0.044086240231990814, -0.018096083775162697, 0.014916148036718369, -0.10755391418933868, -0.024057934060692787, -0.041581664234399796, -0.009493731893599033, -0.21414731442928314, -0.07849729806184769, -0.015424558892846107, 0.06875832378864288, 0.08187821507453918, 0.05718209221959114, 0.09485828876495361, 0.03669736161828041, -0.049296051263809204, 0.06416794657707214, -0.00754922442138195, 0.04546898975968361, -0.09690260142087936, -0.22061537206172943, 0.0024021335411816835, -0.05333106219768524, 0.1042231023311615, -0.19035997986793518, 0.006865684408694506, -0.02629651315510273, 0.1117793619632721, 0.0424678809940815, -0.010236324742436409, -0.04147119075059891, 0.049012795090675354, -0.01619524322450161, -0.06992616504430771, 0.05940898880362511, 0.033996570855379105, -0.11077055335044861, -0.04654320329427719, -0.1170177161693573, 0.02901650406420231, 0.1434568464756012, -0.11821198463439941, -0.14705245196819305, 0.041331417858600616, -0.030671251937747, -0.029653385281562805, -0.036436423659324646, 0.05350514128804207, 0.1851060837507248, -0.015176558867096901, 0.12928883731365204, -0.04046165943145752, -0.019961295649409294, 0.04571545124053955, -0.04061749204993248, 0.003383125877007842, 0.019821124151349068, 0.09026925265789032, -0.1267983466386795, 0.07345159351825714, 0.10105733573436737, -0.05922378972172737, 0.1337953507900238, -0.023656068369746208, -0.04444126784801483, -0.04217902570962906, -0.029521649703383446, 0.018470533192157745, 0.12587352097034454, -0.16553638875484467, 0.00012010036152787507, 0.011673091910779476, 0.029747173190116882, 0.03179246932268143, -0.1881219744682312, -0.008863899856805801, 0.028736485168337822, 0.003950720652937889, -0.04199576750397682, 0.017559070140123367, -0.02186502143740654, 0.08963800221681595, 0.03929460048675537, -0.017312025651335716, 0.07127886265516281, -0.0033898046240210533, -0.08090372383594513, 0.1986260563135147, -0.12869958579540253, -0.07700108736753464, -0.05348629131913185, -0.027290228754281998, -0.027031544595956802, 0.007087267003953457, 0.03445608168840408, -0.0930643379688263, -0.0598209872841835, -0.06402081251144409, 0.010264072567224503, -0.029330801218748093, 0.018636589869856834, 0.03453637659549713, 0.0033246001694351435, 0.09074893593788147, -0.1151224672794342, -0.022919470444321632, -0.06232171878218651, -0.11142611503601074, 0.032245226204395294, -0.00785548985004425, 0.08331016451120377, 0.11973319947719574, -0.0454779714345932, 0.021788710728287697, -0.03744513541460037, 0.24903923273086548, -0.06939709186553955, -0.01082675438374281, 0.08237230777740479, 0.007164554670453072, 0.006016806699335575, 0.09999788552522659, 0.05576392635703087, -0.12816426157951355, 0.05723324045538902, 0.054082050919532776, -0.06771686673164368, -0.1874181628227234, -0.053648076951503754, -0.03941758722066879, -0.11903076618909836, 0.034119293093681335, 0.02829538658261299, 0.09306098520755768, 0.07359614968299866, 0.09359817206859589, 0.09133508056402206, -0.0365791879594326, 0.055774349719285965, 0.08260060101747513, 0.050707872956991196, 0.09719616919755936, -0.05279991403222084, -0.040153052657842636, 0.05617780610918999, -0.05132075399160385, 0.2767595648765564, 0.04260958358645439, 0.03474971279501915, 0.10629509389400482, 0.09070178121328354, -0.011402607895433903, 0.08928689360618591, 0.013765688985586166, -0.06558340787887573, -0.01188377384096384, -0.08096016198396683, -0.03506666421890259, 0.028185872361063957, -0.09906931221485138, 0.04322950541973114, -0.10288511961698532, 0.02279946766793728, 0.0471947081387043, 0.22864001989364624, 0.011041603982448578, -0.319863885641098, -0.09283382445573807, -0.027886513620615005, 0.009381728246808052, -0.03325938060879707, -0.004317876882851124, 0.056000906974077225, -0.09046913683414459, 0.07565180957317352, -0.059231068938970566, 0.09095022082328796, 0.04584232717752457, 0.04477870464324951, 0.026791606098413467, 0.11116259545087814, -0.017679518088698387, 0.052687641233205795, -0.26426005363464355, 0.27648288011550903, 0.021198812872171402, 0.16028741002082825, -0.09169618785381317, -0.006471180822700262, 0.03581841289997101, 0.1304389089345932, 0.1360345035791397, -0.02579670399427414, -0.09111698716878891, -0.1254269927740097, 0.01582307741045952, 0.028076400980353355, 0.1030299961566925, 0.056231603026390076, 0.09132789075374603, -0.015437169000506401, 0.0032236201222985983, 0.09725478291511536, 0.009972617961466312, -0.2071375846862793, -0.08505095541477203, -0.01772911101579666, 0.042284585535526276, -0.07599487155675888, -0.04716939479112625, -0.0882946029305458, -0.016690827906131744, 0.15099811553955078, 0.06576435267925262, -0.02263585850596428, -0.1463051438331604, 0.0786670669913292, 0.07683299481868744, -0.01992383599281311, 0.03329760581254959, 0.0016280219424515963, 0.0633002370595932, 0.04614691808819771, -0.1283882111310959, 0.12912170588970184, -0.08176705986261368, -0.10494384914636612, -0.06893277168273926, 0.019095977768301964, 0.08641707897186279, 0.04833877086639404, 0.024832474067807198, 0.0434793196618557, 0.01491883397102356, -0.08296763896942139, 0.005434195511043072, 0.01687612012028694, 0.056929707527160645, 0.053577158600091934, -0.0631842240691185, -0.025555146858096123, 0.013997839763760567, 0.01572226546704769, 0.108181893825531, 0.18694350123405457, -0.09025915712118149, 0.08682361990213394, 0.06464897841215134, -0.09816090762615204, -0.23965053260326385, 0.15704341232776642, 0.03904738277196884, 0.017732596024870872, 0.06989098340272903, -0.16272969543933868, 0.14615701138973236, 0.05149708688259125, -0.00984110590070486, 0.044579047709703445, -0.30071955919265747, -0.1275368481874466, 0.08571990579366684, 0.1352323293685913, 0.18338152766227722, -0.12126746773719788, -0.020246034488081932, -0.07058554887771606, -0.1172267347574234, 0.1417618691921234, -0.2232893854379654, 0.10532406717538834, 0.020278990268707275, 0.07901522517204285, 0.014869170263409615, -0.04032747820019722, 0.10956777632236481, 0.009019579738378525, 0.12592455744743347, -0.0839599072933197, 0.0012621202040463686, 0.13670197129249573, -0.016191750764846802, 0.05025181919336319, 0.0038176639936864376, 0.04055321589112282, 0.011691232211887836, -0.021468086168169975, -0.07472037523984909, 0.07586108148097992, -0.03340494632720947, -0.07786711305379868, -0.04685915261507034, 0.047678712755441666, 0.06276310980319977, -0.0520755909383297, 0.040236763656139374, -0.016973916441202164, 0.15269212424755096, 0.12568554282188416, 0.14103306829929352, -0.08009852468967438, -0.009161533787846565, 0.06605805456638336, -0.0341569185256958, 0.08203506469726562, -0.1534155309200287, 0.011228461749851704, 0.10093586891889572, 0.0264875590801239, 0.12451814115047455, 0.09810644388198853, -0.06026827543973923, 0.02997172437608242, 0.059940584003925323, -0.09764965623617172, -0.11815091967582703, 0.024036545306444168, -0.043771930038928986, -0.04963644593954086, 0.06200599670410156, 0.12329163402318954, -0.09156910330057144, 0.035789716988801956, -0.016546832397580147, -0.01853039301931858, -0.10382068157196045, 0.15195052325725555, 0.015209428034722805, 0.019062723964452744, -0.08595322072505951, 0.09758882224559784, 0.032255906611680984, -0.06635691970586777, 0.06428119540214539, 0.02252819389104843, -0.10167565196752548, -0.04546792060136795, 0.09104284644126892, 0.25503137707710266, -0.0842084288597107, -0.04332997277379036, -0.09612254798412323, -0.10453531891107559, 0.00812666118144989, 0.20529188215732574, 0.07987550646066666, -0.0054028709419071674, -0.058078162372112274, 0.04199539124965668, -0.15843895077705383, 0.024031830951571465, 0.012615414336323738, 0.03834735229611397, -0.10157165676355362, 0.12597091495990753, 0.010652533732354641, 0.039396483451128006, -0.07658876478672028, 0.0011107055470347404, -0.1269065886735916, 0.013977997936308384, -0.17884530127048492, -0.01716405153274536, -0.04262857884168625, -0.02364286780357361, 0.03508521988987923, -0.03151572495698929, -0.06056378036737442, 0.04161036014556885, -0.08600547909736633, -0.015114115551114082, 0.05542770028114319, 0.018336744979023933, -0.10096735507249832, -0.022666100412607193, 0.001525325234979391, -0.0402345135807991, 0.022064993157982826, 0.06584864854812622, -0.020456958562135696, 0.07218139618635178, -0.21238401532173157, -0.02935902029275894, 0.043038249015808105, -0.014598554000258446, 0.11770524829626083, -0.03398581221699715, -0.026703517884016037, 0.0065027340315282345, 0.09777355939149857, 0.027240091934800148, 0.06323346495628357, -0.10006820410490036, -0.03811461478471756, -0.022493941709399223, -0.04243945702910423, -0.04380088672041893, 0.03608693182468414, 0.11459322273731232, 0.028426161035895348, 0.16878822445869446, -0.11158312112092972, 0.024266643449664116, -0.13190771639347076, -0.03365541994571686, -0.002689021173864603, -0.03692841902375221, -0.04912594333291054, -0.05489582195878029, 0.0766230896115303, -0.07067184895277023, 0.16899965703487396, 0.045690715312957764, 0.11866465955972672, 0.039566535502672195, -0.03305351734161377, -0.037153515964746475, 0.01568424701690674, 0.23226825892925262, 0.05185084417462349, -0.013269723393023014, 0.02402389608323574, 0.07783026248216629, 0.05647354573011398, 0.016305232420563698, 0.2229008823633194, 0.10792315006256104, -0.057398974895477295, 0.09856090694665909, 0.06072724610567093, -0.031110817566514015, -0.08628523349761963, 0.08726415038108826, -0.04838003218173981, 0.11142843961715698, -0.06894878298044205, 0.030308200046420097, 0.06574991345405579, -0.11099778860807419, 0.053596846759319305, -0.08678833395242691, -0.08038900047540665, -0.12225867807865143, -0.04136591777205467, -0.08229228109121323, -0.14658485352993011, 0.010413646697998047, -0.08291124552488327, 0.003190969116985798, 0.07844908535480499, 0.013079696334898472, -0.008638492785394192, 0.21412307024002075, -0.09445716440677643, 0.03522256016731262, 0.05854625254869461, -0.02264781855046749, -0.0387275330722332, -0.10744484513998032, -0.015135321766138077, 0.0189847182482481, 0.005387523211538792, 0.005327909253537655, -0.0200046394020319, -0.02954881265759468, 0.0295571219176054, -0.01626228727400303, -0.08506996929645538, 0.0303071029484272, 0.04425709694623947, 0.004105384927242994, -0.018099555745720863, 0.04180382192134857, -0.01334269531071186, -0.031908851116895676, 0.21521390974521637, -0.0905333161354065, -0.12787708640098572, -0.15609554946422577, 0.2793094515800476, 0.0191921629011631, 0.019939439371228218, 0.016073724254965782, -0.09076521545648575, -0.017549477517604828, 0.24421948194503784, 0.2189735472202301, -0.09991919249296188, 0.0006955421413294971, -0.001873238943517208, -0.013326110318303108, -0.07818160206079483, 0.15378589928150177, 0.05449153110384941, 0.026735274121165276, -0.06938154995441437, -0.022557705640792847, -0.03421739861369133, -0.0409756675362587, -0.05874445661902428, -0.021048177033662796, 0.05995313823223114, 0.011332053691148758, -0.025179442018270493, 0.09542519599199295, -0.0591108538210392, -0.1941870003938675, 0.11319588869810104, -0.15180452167987823, -0.12215021252632141, -0.036240290850400925, 0.027605090290308, 0.008068685419857502, 0.08675588667392731, -0.052006796002388, 0.009207580238580704, 0.10644645243883133, -0.020561331883072853, -0.046095915138721466, -0.0815238282084465, 0.08205468207597733, -0.06300760060548782, 0.22805798053741455, -0.018661247566342354, 0.06983643770217896, 0.10806839913129807, 0.04515143111348152, -0.07961513847112656, 0.09633738547563553, 0.011384934186935425, -0.05635673925280571, 0.024724015966057777, 0.09336752444505692, -0.047242969274520874, 0.04152459278702736, 0.016516776755452156, -0.18745176494121552, 0.04947851970791817, -0.0691298395395279, -0.04397767782211304, -0.06743640452623367, -0.012063145637512207, -0.08254748582839966, 0.13631832599639893, 0.19084668159484863, -0.01179977785795927, 0.03976687788963318, -0.06813997775316238, 0.05548171326518059, 0.06428473442792892, 0.029543336480855942, -0.10116516053676605, -0.20120453834533691, 0.027365989983081818, 0.08321534842252731, -0.023652777075767517, -0.28550392389297485, -0.08108003437519073, 0.008232044987380505, -0.048928141593933105, -0.03218385949730873, 0.054983317852020264, 0.12169555574655533, 0.057348575443029404, -0.05734511837363243, -0.128643199801445, -0.030809687450528145, 0.14519667625427246, -0.11101898550987244, -0.07950282096862793 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # This model is a fine-tuned version of [./checkpoint-10500](https://huggingface.co/./checkpoint-10500) on the COMMON_VOICE - TR dataset. It achieves the following results on the evaluation set: - Loss: 0.7540 - Wer: 0.4647 - Cer: 0.1318 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.999,0.9999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 120.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Cer | Validation Loss | Wer | |:-------------:|:------:|:-----:|:------:|:---------------:|:------:| | 1.0779 | 4.59 | 500 | 0.2354 | 0.8260 | 0.7395 | | 0.7573 | 9.17 | 1000 | 0.2100 | 0.7544 | 0.6960 | | 0.8225 | 13.76 | 1500 | 0.2021 | 0.6867 | 0.6672 | | 0.621 | 18.35 | 2000 | 0.1874 | 0.6824 | 0.6209 | | 0.6362 | 22.94 | 2500 | 0.1904 | 0.6712 | 0.6286 | | 0.624 | 27.52 | 3000 | 0.1820 | 0.6940 | 0.6116 | | 0.4781 | 32.11 | 3500 | 0.1735 | 0.6966 | 0.5989 | | 0.5685 | 36.7 | 4000 | 0.1769 | 0.6742 | 0.5971 | | 0.4384 | 41.28 | 4500 | 0.1767 | 0.6904 | 0.5999 | | 0.5509 | 45.87 | 5000 | 0.1692 | 0.6734 | 0.5641 | | 0.3665 | 50.46 | 5500 | 0.1680 | 0.7018 | 0.5662 | | 0.3914 | 55.05 | 6000 | 0.1631 | 0.7121 | 0.5552 | | 0.2467 | 59.63 | 6500 | 0.1563 | 0.6657 | 0.5374 | | 0.2576 | 64.22 | 7000 | 0.1554 | 0.6920 | 0.5316 | | 0.2711 | 68.81 | 7500 | 0.1495 | 0.6900 | 0.5176 | | 0.2626 | 73.39 | 8000 | 0.1454 | 0.6843 | 0.5043 | | 0.1377 | 77.98 | 8500 | 0.1470 | 0.7383 | 0.5101 | | 0.2005 | 82.57 | 9000 | 0.1430 | 0.7228 | 0.5045 | | 0.1355 | 87.16 | 9500 | 0.1375 | 0.7231 | 0.4869 | | 0.0431 | 91.74 | 10000 | 0.1350 | 0.7397 | 0.4749 | | 0.0586 | 96.33 | 10500 | 0.1339 | 0.7360 | 0.4754 | | 0.0896 | 100.92 | 11000 | 0.7187 | 0.4885 | 0.1398 | | 0.183 | 105.5 | 11500 | 0.7310 | 0.4838 | 0.1392 | | 0.0963 | 110.09 | 12000 | 0.7643 | 0.4759 | 0.1362 | | 0.0437 | 114.68 | 12500 | 0.7525 | 0.4641 | 0.1328 | | 0.1122 | 119.27 | 13000 | 0.7535 | 0.4651 | 0.1317 | ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2.dev0 - Tokenizers 0.11.0
{"language": ["tr"], "tags": ["automatic-speech-recognition", "common_voice", "generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "", "results": []}]}
automatic-speech-recognition
Baybars/wav2vec2-xls-r-1b-turkish
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "common_voice", "generated_from_trainer", "tr", "dataset:common_voice", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[ "tr" ]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #common_voice #generated_from_trainer #tr #dataset-common_voice #endpoints_compatible #region-us
This model is a fine-tuned version of ./checkpoint-10500 on the COMMON\_VOICE - TR dataset. It achieves the following results on the evaluation set: * Loss: 0.7540 * Wer: 0.4647 * Cer: 0.1318 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0002 * train\_batch\_size: 32 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.999,0.9999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 120.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.17.0.dev0 * Pytorch 1.10.2+cu102 * Datasets 1.18.2.dev0 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.999,0.9999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 120.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #common_voice #generated_from_trainer #tr #dataset-common_voice #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.999,0.9999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 120.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ 65, 115, 4, 39 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #common_voice #generated_from_trainer #tr #dataset-common_voice #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0002\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.999,0.9999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 120.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ -0.10841832309961319, 0.07187148928642273, -0.004234189633280039, 0.04809180647134781, 0.12215203791856766, -0.014956892468035221, 0.13339826464653015, 0.1397164762020111, -0.11811815202236176, 0.08549298346042633, 0.09654301404953003, 0.12051492184400558, 0.036838091909885406, 0.09164761751890182, -0.05152159556746483, -0.28048810362815857, 0.038986943662166595, 0.042595840990543365, 0.013027905486524105, 0.10674896091222763, 0.08386146277189255, -0.12828528881072998, 0.04195106029510498, 0.044412460178136826, -0.16471509635448456, 0.017181264236569405, 0.02758469246327877, -0.09780643880367279, 0.12026450783014297, 0.019561169669032097, 0.0982118770480156, 0.03493722528219223, 0.06734880059957504, -0.20676136016845703, 0.010306370444595814, 0.05026440694928169, 0.027672510594129562, 0.06741410493850708, 0.055906664580106735, -0.030186474323272705, 0.11143603920936584, -0.06741318851709366, 0.06446150690317154, 0.04065641760826111, -0.1084822341799736, -0.27902165055274963, -0.07351556420326233, 0.01789569854736328, 0.0613105408847332, 0.09904105216264725, -0.03099391795694828, 0.1474199742078781, -0.04253444820642471, 0.12257442623376846, 0.24188485741615295, -0.2736055850982666, -0.045708443969488144, -0.04780620336532593, 0.04504222050309181, 0.06149878725409508, -0.0908200740814209, -0.009758363477885723, 0.03504125773906708, 0.058609094470739365, 0.10115980356931686, -0.023207655176520348, -0.0743822455406189, -0.023412102833390236, -0.13646449148654938, -0.03566792234778404, 0.15084406733512878, 0.024480445310473442, -0.035172685980796814, -0.07162243127822876, -0.05892433598637581, -0.19375163316726685, -0.04071773216128349, -0.03534078225493431, 0.027089402079582214, -0.036486901342868805, -0.10576945543289185, -0.012427479028701782, -0.08402732759714127, -0.07258392125368118, -0.056793808937072754, 0.18356367945671082, 0.04136229678988457, -0.0032719108276069164, -0.04379855841398239, 0.047707583755254745, -0.048949673771858215, -0.1247064471244812, -0.006881998386234045, 0.05273735150694847, -0.011827782727777958, -0.0237707681953907, -0.05289172753691673, -0.1471402496099472, 0.01679989881813526, 0.10443972051143646, -0.08757638931274414, 0.0904063954949379, -0.059144169092178345, 0.027646876871585846, -0.07013798505067825, 0.1844651699066162, -0.03602705895900726, 0.01850287802517414, 0.0004798955051228404, 0.04670090973377228, 0.039746224880218506, -0.03642640262842178, -0.09445338696241379, 0.0195863526314497, 0.12674283981323242, 0.012538405135273933, -0.05971010401844978, 0.050900254398584366, -0.03849280625581741, -0.013108252547681332, -0.028469542041420937, -0.12146742641925812, 0.03918900340795517, 0.02628546766936779, -0.03300413116812706, 0.045190539211034775, 0.011804967187345028, 0.016451627016067505, -0.06121159717440605, 0.08490265160799026, -0.053334206342697144, 0.033742744475603104, -0.03645244613289833, -0.1439460963010788, 0.029730258509516716, -0.10320035368204117, 0.0020321360789239407, -0.08470693230628967, -0.08159404993057251, -0.02423393540084362, 0.02098993770778179, -0.028042325749993324, -0.02640416845679283, -0.09126070141792297, -0.09586907923221588, 0.04764906316995621, -0.031807735562324524, 0.046580757945775986, -0.06510572135448456, 0.08723098039627075, 0.05185018479824066, 0.07194488495588303, -0.03851989656686783, 0.06703632324934006, -0.04754827544093132, 0.01365689467638731, -0.1847286820411682, 0.09326182305812836, -0.07054345309734344, 0.014802366495132446, -0.09726882725954056, -0.11402656883001328, -0.006356872618198395, 0.007011037785559893, 0.10164790600538254, 0.07386957108974457, -0.1912030726671219, -0.10889405012130737, 0.19755958020687103, -0.10411525517702103, -0.061972904950380325, 0.15196916460990906, -0.0380866676568985, -0.009877619333565235, 0.08414992690086365, 0.2702707052230835, 0.04936203360557556, -0.13155272603034973, -0.020765572786331177, -0.05946268141269684, 0.04732883721590042, -0.006838367320597172, 0.04752520099282265, -0.040622107684612274, 0.0692373663187027, -0.004944511689245701, 0.052122533321380615, 0.024075021967291832, -0.08996644616127014, -0.06383907049894333, -0.03877252712845802, -0.07309471815824509, 0.022874852642416954, 0.042457181960344315, 0.04478762298822403, -0.1251833289861679, -0.10220839828252792, 0.06720878928899765, 0.07403305172920227, -0.11964981257915497, 0.0745839849114418, -0.12895508110523224, 0.08520875126123428, -0.06185206398367882, -0.024842029437422752, -0.19021178781986237, 0.05728134885430336, 0.030369989573955536, -0.002119818702340126, 0.05523970350623131, -0.03568277880549431, 0.07629719376564026, 0.0412987545132637, -0.042374420911073685, -0.03491954877972603, -0.008532409556210041, 0.018257737159729004, -0.08993425220251083, -0.20507627725601196, -0.020090915262699127, -0.04270438849925995, 0.06828153133392334, -0.13777729868888855, 0.016858138144016266, 0.09828074276447296, 0.09187109768390656, 0.040789518505334854, -0.04229944944381714, 0.00005947411045781337, 0.09731414169073105, -0.024171970784664154, -0.043699365109205246, 0.034232985228300095, 0.017715763300657272, -0.0959857627749443, 0.020570650696754456, -0.18944048881530762, 0.10358254611492157, 0.14084172248840332, -0.07465727627277374, -0.05510001257061958, 0.03023814782500267, -0.03144645690917969, -0.030589831992983818, -0.018656345084309578, -0.007514351047575474, 0.2318209558725357, 0.0036674910224974155, 0.13482430577278137, -0.08537236601114273, -0.022213200107216835, 0.057310134172439575, -0.024209080263972282, -0.005247513297945261, 0.10041704028844833, 0.02227947674691677, -0.03803938627243042, 0.09725860506296158, 0.046768054366111755, -0.0923059731721878, 0.15981777012348175, -0.05772215873003006, -0.07803992182016373, -0.032171111553907394, -0.02173585072159767, 0.021804096177220345, 0.08992192894220352, -0.20259837806224823, -0.04161536321043968, 0.017701279371976852, 0.04312003776431084, 0.023331642150878906, -0.213846355676651, 0.0231221504509449, 0.039323318749666214, -0.08649551868438721, -0.059244800359010696, 0.00812490563839674, 0.011102416552603245, 0.08898352086544037, -0.009627968072891235, -0.08249637484550476, 0.009251974523067474, -0.019377456977963448, -0.08381859958171844, 0.14821048080921173, -0.11476775258779526, -0.16049404442310333, -0.10725309699773788, -0.10079427063465118, -0.03303178399801254, 0.02307368442416191, 0.09212794899940491, -0.10278211534023285, -0.04154258221387863, -0.08379822969436646, 0.021693609654903412, -0.03208611533045769, 0.044430430978536606, 0.0413006991147995, -0.025460002943873405, 0.07238463312387466, -0.1243685632944107, -0.022585120052099228, -0.047533463686704636, -0.014063076116144657, 0.030471235513687134, 0.04070713371038437, 0.10207007825374603, 0.1699109524488449, -0.012325615622103214, 0.04125669598579407, -0.04631385952234268, 0.20301522314548492, -0.09284774959087372, -0.0391833633184433, 0.11387471854686737, -0.029968125745654106, 0.04860330745577812, 0.12249460071325302, 0.04784218966960907, -0.08964146673679352, -0.019961265847086906, 0.02092767506837845, -0.04728785157203674, -0.23361586034297943, -0.042102932929992676, -0.0501193106174469, -0.0061261155642569065, 0.05817800015211105, 0.026388846337795258, 0.052565786987543106, 0.00859601330012083, 0.02790077030658722, 0.025391196832060814, 0.00984027050435543, 0.07538217306137085, 0.16857948899269104, 0.03485669195652008, 0.13126203417778015, -0.037064164876937866, -0.044407330453395844, 0.02832917869091034, -0.019710589200258255, 0.20736950635910034, 0.045701880007982254, 0.18437953293323517, 0.046844709664583206, 0.1440582424402237, 0.03139209374785423, 0.06029848754405975, -0.007163538597524166, -0.010185565799474716, 0.017363714054226875, -0.04693372920155525, -0.04168681427836418, 0.009754648432135582, 0.04237502068281174, 0.034655869007110596, -0.11102019995450974, 0.016930866986513138, 0.03498408943414688, 0.30533501505851746, 0.05942150205373764, -0.31377628445625305, -0.09996291249990463, -0.012658857740461826, -0.08045519143342972, -0.01993638649582863, 0.049357786774635315, 0.11954149603843689, -0.051543451845645905, 0.07593832910060883, -0.049552273005247116, 0.07878893613815308, -0.07637355476617813, 0.05236044153571129, 0.01757792755961418, 0.07582596689462662, -0.008881895802915096, 0.024085691198706627, -0.2656862139701843, 0.27600401639938354, 0.02596895769238472, 0.09738266468048096, -0.05218717083334923, 0.00977280642837286, 0.012019017711281776, 0.011394000612199306, 0.06485814601182938, -0.021230453625321388, -0.10359670221805573, -0.1630014330148697, -0.07608967274427414, 0.01723765954375267, 0.13737960159778595, 0.04395260661840439, 0.10641873627901077, -0.011361452750861645, -0.009314147755503654, 0.06496645510196686, -0.08498375117778778, -0.08127404749393463, -0.08421125262975693, 0.007011589594185352, 0.13033680617809296, 0.027092190459370613, -0.04931878298521042, -0.1043199896812439, -0.09692198038101196, 0.12303437292575836, -0.057766567915678024, -0.024486785754561424, -0.10124955326318741, 0.035378340631723404, 0.10069414228200912, -0.07489661872386932, 0.03890598192811012, 0.03515703231096268, 0.09947215020656586, 0.017267702147364616, -0.04358973726630211, 0.1199316754937172, -0.052847858518362045, -0.15951697528362274, -0.0409003347158432, 0.17546145617961884, 0.04033450409770012, 0.08006764948368073, -0.005417947191745043, 0.03540274128317833, -0.00275384820997715, -0.0492396205663681, 0.059335771948099136, 0.032691583037376404, 0.01473928801715374, 0.02061130665242672, 0.004737157840281725, -0.036578912287950516, -0.09795689582824707, -0.015930749475955963, 0.2111532986164093, 0.2355021983385086, -0.07187014073133469, 0.07040055096149445, 0.06235755607485771, -0.056380752474069595, -0.1621549278497696, 0.017823142930865288, 0.07624801993370056, 0.025299564003944397, -0.023878183215856552, -0.18139109015464783, 0.04045524075627327, 0.06077883392572403, -0.005457489285618067, 0.051754143089056015, -0.2896749973297119, -0.14208374917507172, 0.12448693811893463, 0.07830890268087387, 0.08156580477952957, -0.12211613357067108, -0.05425838381052017, -0.0276595801115036, -0.09283995628356934, 0.06759174913167953, -0.0973377376794815, 0.13490892946720123, 0.015353748574852943, 0.11220446974039078, 0.022330516949295998, -0.04339785501360893, 0.09381771832704544, 0.021228710189461708, 0.053890589624643326, -0.04577501863241196, 0.010142462328076363, 0.041697658598423004, -0.03326571360230446, 0.045278798788785934, -0.048492059111595154, 0.015297307632863522, -0.06272324174642563, -0.04286303371191025, -0.08251006156206131, 0.007238396443426609, 0.004618851467967033, -0.033670682460069656, -0.0327962227165699, -0.026619069278240204, 0.07040630280971527, -0.003984764218330383, 0.16349883377552032, -0.0430874228477478, 0.14154212176799774, 0.1460903435945511, 0.1494060903787613, -0.10728129744529724, -0.051567982882261276, 0.013126800768077374, -0.04038020595908165, 0.05803299322724342, -0.10360725969076157, 0.05591592937707901, 0.12895166873931885, 0.03235558047890663, 0.14509251713752747, 0.07430137693881989, -0.07797978818416595, 0.054470472037792206, 0.035361386835575104, -0.12800636887550354, -0.15737703442573547, 0.004657707642763853, -0.005546670872718096, -0.07995054125785828, 0.07927591353654861, 0.14015734195709229, -0.058238789439201355, -0.005734138190746307, -0.004851475358009338, 0.0011557332472875714, -0.06476176530122757, 0.20892538130283356, 0.02653372846543789, 0.044454868882894516, -0.11761708557605743, 0.06937579065561295, 0.03208460658788681, -0.14453302323818207, 0.07987476885318756, 0.06657391041517258, -0.061823129653930664, -0.02004561759531498, 0.013281917199492455, 0.14427201449871063, -0.00007716842810623348, -0.0746060237288475, -0.1323540210723877, -0.1347239464521408, 0.09583603590726852, 0.21322211623191833, 0.04644075781106949, 0.02112540602684021, -0.0835333839058876, 0.034285593777894974, -0.1443353295326233, 0.07544437050819397, 0.05806480348110199, 0.05409741401672363, -0.14451897144317627, 0.1758088618516922, 0.005036806687712669, 0.02165529876947403, -0.02030576579272747, -0.019299060106277466, -0.0774914100766182, 0.03840133920311928, -0.13992930948734283, -0.000844302645418793, -0.04688918590545654, -0.0031769308261573315, 0.023783834651112556, -0.058836713433265686, -0.07408097386360168, 0.039474599063396454, -0.11323905736207962, -0.011424534022808075, 0.013657460920512676, 0.02658216655254364, -0.11298888176679611, -0.006182570476084948, 0.012405795976519585, -0.07648167759180069, 0.09050280600786209, 0.10069417208433151, -0.04562174528837204, 0.06637709587812424, -0.10147138684988022, -0.03176968917250633, 0.09890513122081757, 0.0012631089193746448, 0.04506098851561546, -0.11418662965297699, -0.024151436984539032, 0.028409156948328018, 0.0434538871049881, 0.019041061401367188, 0.10072997957468033, -0.09320689737796783, 0.014086260460317135, -0.04219219461083412, -0.05748556926846504, -0.06525596231222153, 0.0392598919570446, 0.10541117191314697, 0.03862467780709267, 0.17971666157245636, -0.10507773607969284, 0.04176729917526245, -0.18121105432510376, 0.016301244497299194, -0.015714818611741066, -0.10364697128534317, -0.07568434625864029, -0.031134171411395073, 0.08536915481090546, -0.07348418980836868, 0.10029609501361847, -0.020698530599474907, 0.0390751026570797, 0.026687296107411385, -0.08600141108036041, -0.06912390887737274, 0.025405628606677055, 0.21427620947360992, 0.04915570095181465, -0.06586099416017532, 0.06644172221422195, -0.00046749506145715714, 0.10939382761716843, 0.18150930106639862, 0.13222099840641022, 0.14013347029685974, 0.05174947530031204, 0.11546772718429565, 0.08676046878099442, -0.041458919644355774, -0.16588015854358673, 0.05658490210771561, -0.062257830053567886, 0.12098285555839539, -0.0009926117490977049, 0.22836430370807648, 0.1049436628818512, -0.12708014249801636, 0.07024969905614853, -0.0335606224834919, -0.09742382168769836, -0.13674353063106537, -0.07563578337430954, -0.09926174581050873, -0.1610516905784607, 0.018260261043906212, -0.12806151807308197, 0.07264554500579834, 0.042902108281850815, 0.03621438890695572, 0.011493412777781487, 0.16097714006900787, -0.015251974575221539, 0.0026141495909541845, 0.09711255133152008, -0.009272660128772259, -0.04925219342112541, -0.06961602717638016, -0.07668952643871307, 0.06981077045202255, -0.006645608693361282, 0.060892254114151, -0.0007034118752926588, -0.056058723479509354, 0.03772615268826485, -0.05473269149661064, -0.1215016320347786, 0.026544468477368355, 0.01652819849550724, 0.07161551713943481, 0.05729223042726517, 0.04290119186043739, -0.037301164120435715, 0.0006874707178212702, 0.17721931636333466, -0.09501950442790985, -0.07393743842840195, -0.11406825482845306, 0.21541525423526764, 0.0412987545132637, 0.008500519208610058, 0.010425982996821404, -0.08100651949644089, -0.017041435465216637, 0.1817621886730194, 0.19045469164848328, 0.0042375195771455765, 0.00486591225489974, -0.017940634861588478, -0.003282509744167328, -0.042184434831142426, 0.0524754673242569, 0.1326950043439865, 0.045014962553977966, -0.03336500748991966, -0.02818906493484974, -0.06485550850629807, -0.01613815687596798, -0.033934928476810455, 0.04538965970277786, 0.012193286791443825, -0.02394055388867855, -0.032412804663181305, 0.07303201407194138, -0.08397375047206879, -0.1089901328086853, -0.00006872180529171601, -0.2127636820077896, -0.13790787756443024, 0.02047099359333515, 0.07925659418106079, 0.04530050605535507, 0.044668227434158325, -0.013174519874155521, -0.008432873524725437, 0.09836678206920624, -0.0027752346359193325, -0.0895485207438469, -0.04668331891298294, 0.07489272952079773, -0.15751148760318756, 0.13649575412273407, -0.03052114136517048, 0.07458508014678955, 0.10862696915864944, 0.09822526574134827, -0.04564663767814636, 0.09683209657669067, 0.033211544156074524, -0.10887768864631653, 0.011889280751347542, 0.20261293649673462, -0.04086223244667053, 0.12284990400075912, 0.045839667320251465, -0.1431095153093338, 0.0064218672923743725, -0.09286589175462723, -0.0495176836848259, -0.032704513520002365, -0.049841683357954025, -0.054113514721393585, 0.1076817512512207, 0.15924221277236938, -0.0353364534676075, 0.0020551609341055155, -0.04484247788786888, -0.007273759692907333, 0.06530561298131943, 0.030709806829690933, -0.06171495094895363, -0.2941707372665405, 0.01093438919633627, 0.02668294496834278, -0.006461154669523239, -0.25530776381492615, -0.09319392591714859, 0.015555215068161488, -0.05699215456843376, -0.08223194628953934, 0.08007823675870895, 0.06782035529613495, 0.04399551451206207, -0.05828024446964264, -0.02757786586880684, -0.016028814017772675, 0.19926202297210693, -0.16757771372795105, -0.06751962751150131 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the COMMON_VOICE - TR dataset. It achieves the following results on the evaluation set: - Loss: 0.4164 - Wer: 0.3098 - Cer: 0.0764 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Language Model N-gram language model is trained by [mpoyraz](https://huggingface.co/mpoyraz/wav2vec2-xls-r-300m-cv7-turkish) on a Turkish Wikipedia articles using KenLM and [ngram-lm-wiki](https://github.com/mpoyraz/ngram-lm-wiki) repo was used to generate arpa LM and convert it into binary format. ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 64 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 100.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | Cer | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:| | 0.6356 | 9.09 | 500 | 0.5055 | 0.5536 | 0.1381 | | 0.3847 | 18.18 | 1000 | 0.4002 | 0.4247 | 0.1065 | | 0.3377 | 27.27 | 1500 | 0.4193 | 0.4167 | 0.1078 | | 0.2175 | 36.36 | 2000 | 0.4351 | 0.3861 | 0.0974 | | 0.2074 | 45.45 | 2500 | 0.3962 | 0.3622 | 0.0916 | | 0.159 | 54.55 | 3000 | 0.4062 | 0.3526 | 0.0888 | | 0.1882 | 63.64 | 3500 | 0.3991 | 0.3445 | 0.0850 | | 0.1766 | 72.73 | 4000 | 0.4214 | 0.3396 | 0.0847 | | 0.116 | 81.82 | 4500 | 0.4182 | 0.3265 | 0.0812 | | 0.0718 | 90.91 | 5000 | 0.4259 | 0.3191 | 0.0781 | | 0.019 | 100.0 | 5500 | 0.4164 | 0.3098 | 0.0764 | ## Evaluation Commands Please install [unicode_tr](https://pypi.org/project/unicode_tr/) package before running evaluation. It is used for Turkish text processing. 1. To evaluate on `mozilla-foundation/common_voice_7_0` with split `test` ```bash python eval.py --model_id Baybars/wav2vec2-xls-r-300m-cv8-turkish --dataset mozilla-foundation/common_voice_8_0 --config tr --split test ``` 2. To evaluate on `speech-recognition-community-v2/dev_data` ```bash python eval.py --model_id Baybars/wav2vec2-xls-r-300m-cv8-turkish --dataset speech-recognition-community-v2/dev_data --config tr --split validation --chunk_length_s 5.0 --stride_length_s 1.0 ``` ### Framework versions - Transformers 4.17.0.dev0 - Pytorch 1.10.2+cu102 - Datasets 1.18.2.dev0 - Tokenizers 0.11.0
{"language": ["tr"], "license": "apache-2.0", "tags": ["automatic-speech-recognition", "common_voice", "generated_from_trainer", "hf-asr-leaderboard", "robust-speech-event", "tr"], "datasets": ["common_voice"], "model-index": [{"name": "", "results": []}]}
automatic-speech-recognition
Baybars/wav2vec2-xls-r-300m-cv8-turkish
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "common_voice", "generated_from_trainer", "hf-asr-leaderboard", "robust-speech-event", "tr", "dataset:common_voice", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[ "tr" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #common_voice #generated_from_trainer #hf-asr-leaderboard #robust-speech-event #tr #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us
This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the COMMON\_VOICE - TR dataset. It achieves the following results on the evaluation set: * Loss: 0.4164 * Wer: 0.3098 * Cer: 0.0764 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Language Model -------------- N-gram language model is trained by mpoyraz on a Turkish Wikipedia articles using KenLM and ngram-lm-wiki repo was used to generate arpa LM and convert it into binary format. Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0005 * train\_batch\_size: 64 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * num\_epochs: 100.0 * mixed\_precision\_training: Native AMP ### Training results Evaluation Commands ------------------- Please install unicode\_tr package before running evaluation. It is used for Turkish text processing. 1. To evaluate on 'mozilla-foundation/common\_voice\_7\_0' with split 'test' 2. To evaluate on 'speech-recognition-community-v2/dev\_data' ### Framework versions * Transformers 4.17.0.dev0 * Pytorch 1.10.2+cu102 * Datasets 1.18.2.dev0 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 100.0\n* mixed\\_precision\\_training: Native AMP", "### Training results\n\n\n\nEvaluation Commands\n-------------------\n\n\nPlease install unicode\\_tr package before running evaluation. It is used for Turkish text processing.\n\n\n1. To evaluate on 'mozilla-foundation/common\\_voice\\_7\\_0' with split 'test'\n2. To evaluate on 'speech-recognition-community-v2/dev\\_data'", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #common_voice #generated_from_trainer #hf-asr-leaderboard #robust-speech-event #tr #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 100.0\n* mixed\\_precision\\_training: Native AMP", "### Training results\n\n\n\nEvaluation Commands\n-------------------\n\n\nPlease install unicode\\_tr package before running evaluation. It is used for Turkish text processing.\n\n\n1. To evaluate on 'mozilla-foundation/common\\_voice\\_7\\_0' with split 'test'\n2. To evaluate on 'speech-recognition-community-v2/dev\\_data'", "### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ 87, 131, 86, 39 ]
[ "passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #common_voice #generated_from_trainer #hf-asr-leaderboard #robust-speech-event #tr #dataset-common_voice #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 100.0\n* mixed\\_precision\\_training: Native AMP### Training results\n\n\n\nEvaluation Commands\n-------------------\n\n\nPlease install unicode\\_tr package before running evaluation. It is used for Turkish text processing.\n\n\n1. To evaluate on 'mozilla-foundation/common\\_voice\\_7\\_0' with split 'test'\n2. To evaluate on 'speech-recognition-community-v2/dev\\_data'### Framework versions\n\n\n* Transformers 4.17.0.dev0\n* Pytorch 1.10.2+cu102\n* Datasets 1.18.2.dev0\n* Tokenizers 0.11.0" ]
[ -0.09730944782495499, 0.05774121358990669, -0.005445437040179968, -0.003405623137950897, 0.10625288635492325, 0.024871716275811195, 0.11452748626470566, 0.15810136497020721, -0.005029147490859032, 0.10494175553321838, 0.05073300749063492, 0.06206858158111572, 0.1061105728149414, 0.09435465931892395, -0.04560180753469467, -0.1980549842119217, 0.03237244859337807, -0.051666706800460815, -0.014647707343101501, 0.13740552961826324, 0.12118242681026459, -0.06350651383399963, 0.006641135085374117, 0.016055798158049583, -0.08673499524593353, 0.0023633651435375214, -0.014842215925455093, -0.07496444135904312, 0.0754849836230278, 0.013841797597706318, 0.06915456056594849, 0.03170747309923172, 0.005328322295099497, -0.2822020351886749, 0.0045725274831056595, 0.062063127756118774, 0.027986755594611168, 0.02956237830221653, 0.13945922255516052, -0.04847777634859085, 0.13629166781902313, -0.07939966768026352, 0.046555791050195694, 0.08242613077163696, -0.0723811537027359, -0.21988672018051147, -0.1258067637681961, 0.03326871991157532, 0.17582100629806519, 0.08707009255886078, -0.059856265783309937, 0.07443412393331528, -0.07921013981103897, 0.10074439644813538, 0.11596131324768066, -0.2412859946489334, -0.053108423948287964, -0.10101902484893799, 0.018689408898353577, 0.061031524091959, -0.06058645248413086, -0.007931671105325222, -0.004587156698107719, 0.015197190456092358, 0.005179230123758316, -0.0032172503415495157, -0.025035779923200607, -0.05290450528264046, -0.1268995702266693, -0.09728322923183441, 0.1692580282688141, 0.0542619414627552, -0.03885743021965027, -0.10380145162343979, -0.0019653018098324537, -0.19587185978889465, -0.016922276467084885, 0.006654731929302216, -0.024924717843532562, -0.026363618671894073, 0.011541374959051609, -0.0027046508621424437, -0.09279102087020874, -0.07193100452423096, 0.0036260446067899466, 0.16300886869430542, 0.060949910432100296, -0.03335431590676308, 0.03455951064825058, 0.11614858359098434, 0.000437240581959486, -0.13606272637844086, -0.05628645420074463, 0.04634575918316841, -0.1538500338792801, -0.02116849273443222, -0.026895854622125626, -0.12198172509670258, 0.049642838537693024, 0.15500761568546295, 0.015316765755414963, 0.11998369544744492, -0.05359378829598427, 0.019680768251419067, -0.09578238427639008, 0.19012963771820068, 0.027302851900458336, -0.02164652571082115, -0.048079393804073334, 0.09092545509338379, -0.0011820964282378554, -0.0009313850896432996, -0.016499504446983337, 0.008677016012370586, 0.1536957174539566, 0.12670472264289856, -0.0011717276647686958, 0.001541240606456995, -0.06362421065568924, -0.010041887871921062, 0.01462804526090622, -0.1485711634159088, 0.03885984420776367, 0.07114919275045395, -0.03802521526813507, -0.014096208848059177, 0.00924716517329216, 0.009981825947761536, -0.0833774209022522, 0.08519555628299713, 0.008827089332044125, 0.02217116393148899, -0.050253529101610184, -0.07803799957036972, 0.02530635893344879, -0.016761260107159615, -0.040196288377046585, -0.10661261528730392, -0.07435709983110428, -0.06252234429121017, 0.027804059907794, -0.05398895591497421, -0.01037549041211605, -0.038843318819999695, -0.0739016979932785, 0.04451916366815567, -0.029630577191710472, 0.15160906314849854, -0.06142475828528404, 0.10030381381511688, 0.04230551794171333, 0.01926730014383793, 0.11184244602918625, 0.07124540954828262, -0.024770798161625862, 0.0425419919192791, -0.11553801596164703, 0.14587444067001343, -0.11044377833604813, 0.037675537168979645, -0.17003095149993896, -0.08915464580059052, 0.009126479737460613, -0.03680701181292534, 0.08347971737384796, 0.09623939543962479, -0.19589543342590332, -0.054535459727048874, 0.13141895830631256, -0.08871718496084213, -0.07252724468708038, 0.14710097014904022, 0.024805651977658272, -0.025110751390457153, 0.05377240478992462, 0.20507733523845673, 0.1383993923664093, -0.0848727896809578, -0.02129490114748478, -0.07117261737585068, 0.04974125325679779, 0.08572568744421005, 0.10105560719966888, -0.06257611513137817, 0.00147019955329597, 0.008946801535785198, -0.09319471567869186, 0.06002868339419365, -0.030000396072864532, -0.06861164420843124, -0.016830619424581528, -0.0498226173222065, 0.011362283490598202, 0.00547434575855732, -0.0019594519399106503, -0.1246667355298996, -0.1545981764793396, -0.03530575707554817, 0.10170455276966095, -0.0619751513004303, 0.030699986964464188, -0.10092807561159134, 0.07807930558919907, 0.008031642995774746, 0.013455605134367943, -0.15724660456180573, 0.016752639785408974, 0.05944167822599411, -0.07701674848794937, 0.025381969287991524, -0.07494513690471649, 0.032031018286943436, 0.011623089201748371, -0.018930254504084587, -0.07115298509597778, -0.02136366255581379, 0.0005246592336334288, -0.05093064159154892, -0.1299804300069809, -0.032166801393032074, -0.012252171523869038, 0.21996110677719116, -0.14346322417259216, 0.027661124244332314, 0.1306295245885849, 0.1120394915342331, -0.015781423076987267, -0.04075733572244644, 0.00879585649818182, 0.05014234781265259, -0.004393822979182005, -0.060486264526844025, 0.00526047870516777, 0.00025916306185536087, -0.10406186431646347, 0.035513654351234436, -0.1806248128414154, -0.09441863000392914, 0.09379034489393234, 0.04814545065164566, -0.0847807452082634, 0.015689430758357048, -0.0602419376373291, -0.04853673651814461, -0.027818337082862854, -0.05578004568815231, 0.17809417843818665, 0.06350488215684891, 0.11649452149868011, -0.06373665481805801, -0.06107499450445175, -0.004198852460831404, 0.0017348716501146555, -0.02764594554901123, 0.12993207573890686, -0.050339795649051666, -0.08410167694091797, 0.05471726506948471, 0.06982841342687607, -0.07612375169992447, 0.08814690262079239, -0.04889349639415741, -0.09446822106838226, -0.03447892144322395, 0.11627910286188126, 0.0608648918569088, 0.03629152476787567, -0.16028594970703125, 0.018300728872418404, 0.04941949620842934, 0.03175017237663269, 0.024280421435832977, -0.1834702044725418, 0.043316975235939026, 0.03390909358859062, -0.09378780424594879, -0.016053523868322372, 0.029569225385785103, 0.015231958590447903, 0.06669638305902481, -0.014516118913888931, -0.006440990604460239, -0.015580357052385807, -0.05716880038380623, -0.11989196389913559, 0.15035352110862732, -0.1175997331738472, -0.19414620101451874, -0.13950613141059875, -0.02084798738360405, -0.028329871594905853, -0.027832815423607826, 0.09751591086387634, -0.11853337287902832, -0.05345296859741211, -0.07776942849159241, 0.0037173002492636442, -0.028534963726997375, 0.01667804643511772, -0.031345706433057785, 0.00913456454873085, 0.07900864630937576, -0.0907081812620163, -0.007167046424001455, 0.015776796266436577, -0.03801046684384346, 0.04060155525803566, 0.028099842369556427, 0.05240583419799805, 0.15353941917419434, -0.005488227121531963, 0.0396491140127182, -0.020025959238409996, 0.1519889086484909, -0.13570062816143036, -0.005211775656789541, 0.1273961216211319, -0.03322816267609596, 0.04646551236510277, 0.14095363020896912, 0.002246275544166565, -0.06566261500120163, 0.020647820085287094, 0.01545119658112526, -0.0067995511926710606, -0.30932939052581787, -0.03547513112425804, -0.07355737686157227, -0.05191831290721893, 0.07137452811002731, 0.05174983665347099, 0.015873219817876816, 0.01383553072810173, -0.058518584817647934, -0.05665266141295433, 0.05899542197585106, 0.06999343633651733, 0.1562436819076538, 0.04273334518074989, 0.09176341444253922, -0.035580333322286606, 0.006790943909436464, 0.03165068477392197, 0.008876263163983822, 0.22433261573314667, 0.014855002984404564, 0.2063964456319809, 0.06911806762218475, 0.13298436999320984, 0.0005404639523476362, 0.023659734055399895, 0.023748410865664482, 0.03471808508038521, 0.05912793427705765, -0.06806199252605438, -0.05003993213176727, 0.025530699640512466, 0.11315721273422241, -0.0001818314049160108, -0.055628929287195206, -0.009638294577598572, 0.0676594153046608, 0.3059404790401459, 0.059731945395469666, -0.22915056347846985, -0.07529187202453613, 0.008119112811982632, -0.12126859277486801, -0.05701567232608795, -0.021615851670503616, 0.07086397707462311, -0.12023039907217026, 0.10336792469024658, -0.05190073698759079, 0.09601246565580368, -0.07133034616708755, -0.018613914027810097, 0.06617056578397751, 0.06352387368679047, 0.029117681086063385, 0.079985111951828, -0.2501089870929718, 0.21838277578353882, -0.005349668674170971, 0.09030334651470184, -0.05145769938826561, 0.062392424792051315, 0.03746041655540466, -0.04342534765601158, 0.06996048986911774, -0.015429571270942688, -0.09034941345453262, -0.12279734015464783, -0.11347299069166183, 0.0010710859205573797, 0.12251526862382889, -0.0385252870619297, 0.09198286384344101, -0.05052994564175606, -0.020601756870746613, -0.014413890428841114, -0.09516987204551697, -0.15311332046985626, -0.10503905266523361, 0.07347068190574646, 0.04084615781903267, 0.10782317072153091, -0.07787902653217316, -0.06879382580518723, -0.07494890689849854, 0.15471966564655304, -0.17201374471187592, -0.06907462328672409, -0.11282844096422195, 0.02065301686525345, 0.15509720146656036, -0.06805090606212616, 0.013099579140543938, -0.013755072839558125, 0.11326491087675095, 0.021525852382183075, 0.02391323633491993, 0.04494683817028999, -0.06800775974988937, -0.1799190491437912, -0.016005657613277435, 0.2236996442079544, -0.004103231243789196, 0.07535399496555328, 0.010700847953557968, 0.01579226553440094, -0.02591860108077526, -0.07014451175928116, 0.034700892865657806, 0.01778189279139042, -0.04974880442023277, 0.10630720853805542, -0.005279852543026209, -0.09932801872491837, -0.16306109726428986, -0.07957632094621658, 0.15592901408672333, 0.24896518886089325, -0.051374662667512894, 0.059702545404434204, -0.02471824921667576, -0.04978039115667343, -0.10326184332370758, -0.022602437064051628, 0.15435732901096344, 0.03245815634727478, -0.004231217782944441, -0.1088506430387497, -0.016665257513523102, 0.08061345666646957, -0.01282427180558443, 0.09429576992988586, -0.2948017418384552, -0.1427319049835205, 0.08140704035758972, 0.05011517554521561, -0.10287590324878693, -0.14980009198188782, -0.10123934596776962, -0.03068733774125576, -0.0449821799993515, 0.01261033583432436, -0.005884973332285881, 0.11739088594913483, 0.002814545063301921, 0.06476825475692749, 0.026871008798480034, -0.0452975258231163, 0.1966307908296585, 0.017244823276996613, -0.005569112952798605, -0.035647351294755936, 0.07247218489646912, 0.019185436889529228, -0.06663896888494492, 0.05767253786325455, -0.10458137094974518, 0.01267124991863966, -0.14047971367835999, -0.01804913580417633, -0.0662616640329361, 0.04577070474624634, -0.04026339575648308, 0.030799750238656998, -0.01179727166891098, 0.006463389378041029, 0.08025386929512024, 0.004387550987303257, 0.029716912657022476, -0.0625971183180809, 0.067771777510643, 0.16054096817970276, 0.10957390815019608, 0.03978034481406212, -0.12276292592287064, -0.00797470472753048, 0.014443457126617432, 0.005367610603570938, -0.10074266046285629, 0.037375036627054214, 0.11884203553199768, 0.02100849710404873, 0.16639859974384308, -0.0018592366250231862, -0.1175718903541565, 0.029243938624858856, 0.07291211187839508, -0.04271506518125534, -0.16878534853458405, -0.0013386317295953631, 0.10430789738893509, -0.10368639975786209, -0.07337098568677902, 0.11683648079633713, -0.029497597366571426, -0.0002881028631236404, 0.02487255074083805, 0.026816345751285553, -0.056785084307193756, 0.2148924171924591, -0.008355575613677502, 0.09583937376737595, -0.07464118301868439, 0.05826381593942642, 0.08361798524856567, -0.10474085062742233, 0.029169192537665367, 0.102955661714077, -0.08329354226589203, -0.04858022555708885, -0.04902171716094017, 0.016635634005069733, 0.05643406882882118, -0.07624682784080505, -0.07185757160186768, -0.14136244356632233, 0.055849816650152206, 0.030251238495111465, 0.0319521427154541, 0.017406495288014412, -0.022570345550775528, 0.002328879199922085, -0.06603282690048218, 0.1187836155295372, 0.09870506078004837, 0.05676659941673279, -0.09232652932405472, 0.10615672171115875, 0.029777145013213158, 0.018329745158553123, 0.01816382259130478, -0.02421867474913597, -0.016871102154254913, 0.04863160848617554, -0.13562461733818054, 0.0070914169773459435, -0.04094093292951584, 0.001911657047457993, 0.03707989677786827, -0.06660282611846924, -0.04913913831114769, 0.039406534284353256, -0.10573233664035797, -0.08902160823345184, -0.053952306509017944, 0.08361198008060455, -0.10744385421276093, -0.016473593190312386, 0.055024176836013794, -0.13242274522781372, 0.11331181228160858, 0.07635288685560226, -0.014531525783240795, 0.03321252763271332, -0.07166258245706558, -0.03873426094651222, 0.02056233026087284, 0.047109995037317276, 0.008811186999082565, -0.21454904973506927, 0.019695300608873367, 0.02418460324406624, -0.009315483272075653, 0.03217785060405731, 0.09466645866632462, -0.12318255752325058, -0.022931579500436783, -0.055141571909189224, -0.024903548881411552, -0.05615156888961792, 0.06276915967464447, 0.04520939663052559, 0.05685805156826973, 0.18075311183929443, -0.08924821019172668, 0.0812939703464508, -0.16588222980499268, 0.005897379480302334, -0.03278908133506775, -0.05643634498119354, -0.05109766498208046, -0.025328435003757477, 0.0927121564745903, -0.05243346095085144, 0.05138574168086052, -0.030629156157374382, 0.11401371657848358, 0.05666203051805496, -0.04831034317612648, -0.005321061238646507, 0.022169683128595352, 0.1811525970697403, 0.04257144778966904, -0.00547673087567091, 0.048436447978019714, -0.09880324453115463, 0.02734503522515297, -0.033235564827919006, 0.11177290976047516, 0.20963811874389648, 0.031671151518821716, 0.07278738915920258, 0.048406872898340225, -0.09168343245983124, -0.13126739859580994, 0.07278712093830109, -0.08624129742383957, 0.11882926523685455, -0.02191421575844288, 0.2254595309495926, 0.1036616638302803, -0.191594198346138, 0.06454381346702576, -0.014961396344006062, -0.09535329788923264, -0.06607082486152649, -0.07701463252305984, -0.06666834652423859, -0.14752520620822906, 0.009272458963096142, -0.055583830922842026, 0.07156625390052795, 0.028092224150896072, 0.04470342397689819, 0.005639804992824793, 0.1506694257259369, -0.01782539300620556, -0.054516248404979706, 0.126205176115036, -0.0331670306622982, -0.009795174933969975, -0.0013788025826215744, -0.064000204205513, 0.05542653426527977, -0.0019280374981462955, 0.09977610409259796, 0.04431384056806564, -0.07616310566663742, 0.03292419761419296, -0.0282598827034235, -0.10558899492025375, 0.039294399321079254, 0.006422512698918581, 0.04632148891687393, 0.1274745613336563, 0.06173553317785263, 0.0005252073751762509, -0.01776142604649067, 0.16653282940387726, -0.05026054009795189, -0.08842305839061737, -0.10860239714384079, 0.14956335723400116, 0.02505224011838436, -0.013549677096307278, 0.0178082175552845, -0.10046102851629257, -0.022852765396237373, 0.16039308905601501, 0.0854513868689537, 0.029302392154932022, -0.00830029882490635, -0.008000067435204983, -0.0009175137965939939, -0.02345805987715721, 0.05714096873998642, 0.10911518335342407, 0.015724902972579002, -0.02518787980079651, 0.05018531531095505, -0.03815858066082001, -0.06683529913425446, -0.05788503959774971, 0.057600270956754684, -0.0265330970287323, -0.03546776995062828, -0.006215633824467659, 0.16353103518486023, -0.016909467056393623, -0.16133157908916473, 0.006721766199916601, -0.11951444298028946, -0.17978353798389435, 0.006411246955394745, 0.12756787240505219, 0.01605166681110859, 0.01823326200246811, -0.005974808242172003, -0.033282265067100525, 0.13780620694160461, -0.01174170058220625, -0.01929059997200966, -0.09574190527200699, 0.06845464557409286, -0.09310528635978699, 0.13221375644207, -0.008382268249988556, 0.10160950571298599, 0.11921973526477814, 0.04498952627182007, -0.09279509633779526, 0.09720518440008163, 0.0944228246808052, -0.10112275928258896, 0.11598402261734009, 0.2145831286907196, -0.03336992487311363, 0.09098935127258301, 0.06219072639942169, -0.08490706980228424, 0.016921034082770348, -0.07640814036130905, -0.0034814795944839716, -0.09608960151672363, 0.013671312481164932, -0.06456486880779266, 0.10397637635469437, 0.223049134016037, -0.06395292282104492, 0.010599520988762379, -0.0529305599629879, -0.017270555719733238, 0.02158411778509617, 0.14360173046588898, -0.02278885431587696, -0.26467931270599365, 0.025387097150087357, 0.03431304916739464, 0.052484750747680664, -0.21385517716407776, -0.08174227178096771, -0.0062084379605948925, -0.05462273955345154, -0.03561178967356682, 0.1396547108888626, 0.03268267214298248, 0.017827700823545456, -0.036822326481342316, -0.17580434679985046, 0.000992265180684626, 0.17519281804561615, -0.16802501678466797, -0.03157868981361389 ]
null
null
transformers
# Query Generation This model is the t5-base model from [docTTTTTquery](https://github.com/castorini/docTTTTTquery). The T5-base model was trained on the [MS MARCO Passage Dataset](https://github.com/microsoft/MSMARCO-Passage-Ranking), which consists of about 500k real search queries from Bing together with the relevant passage. The model can be used for query generation to learn semantic search models without requiring annotated training data: [Synthetic Query Generation](https://github.com/UKPLab/sentence-transformers/tree/master/examples/unsupervised_learning/query_generation). ## Usage ```python from transformers import T5Tokenizer, T5ForConditionalGeneration tokenizer = T5Tokenizer.from_pretrained('model-name') model = T5ForConditionalGeneration.from_pretrained('model-name') para = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects." input_ids = tokenizer.encode(para, return_tensors='pt') outputs = model.generate( input_ids=input_ids, max_length=64, do_sample=True, top_p=0.95, num_return_sequences=3) print("Paragraph:") print(para) print("\nGenerated Queries:") for i in range(len(outputs)): query = tokenizer.decode(outputs[i], skip_special_tokens=True) print(f'{i + 1}: {query}') ```
{}
text2text-generation
BeIR/query-gen-msmarco-t5-base-v1
[ "transformers", "pytorch", "jax", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #jax #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Query Generation This model is the t5-base model from docTTTTTquery. The T5-base model was trained on the MS MARCO Passage Dataset, which consists of about 500k real search queries from Bing together with the relevant passage. The model can be used for query generation to learn semantic search models without requiring annotated training data: Synthetic Query Generation. ## Usage
[ "# Query Generation\nThis model is the t5-base model from docTTTTTquery.\n\nThe T5-base model was trained on the MS MARCO Passage Dataset, which consists of about 500k real search queries from Bing together with the relevant passage.\n\nThe model can be used for query generation to learn semantic search models without requiring annotated training data: Synthetic Query Generation.", "## Usage" ]
[ "TAGS\n#transformers #pytorch #jax #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Query Generation\nThis model is the t5-base model from docTTTTTquery.\n\nThe T5-base model was trained on the MS MARCO Passage Dataset, which consists of about 500k real search queries from Bing together with the relevant passage.\n\nThe model can be used for query generation to learn semantic search models without requiring annotated training data: Synthetic Query Generation.", "## Usage" ]
[ 51, 87, 3 ]
[ "passage: TAGS\n#transformers #pytorch #jax #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Query Generation\nThis model is the t5-base model from docTTTTTquery.\n\nThe T5-base model was trained on the MS MARCO Passage Dataset, which consists of about 500k real search queries from Bing together with the relevant passage.\n\nThe model can be used for query generation to learn semantic search models without requiring annotated training data: Synthetic Query Generation.## Usage" ]
[ 0.0035321186296641827, 0.07132449746131897, -0.0034816828556358814, 0.05618482828140259, 0.16366024315357208, 0.0070151109248399734, 0.05777270719408989, 0.09940236061811447, 0.03087661601603031, -0.0799681693315506, 0.09968629479408264, 0.09432600438594818, -0.02423050068318844, 0.1521737277507782, -0.04812584072351456, -0.24514101445674896, 0.00973388273268938, 0.02144435979425907, -0.05103942006826401, 0.11938455700874329, 0.09747091680765152, -0.05866660177707672, 0.13916876912117004, -0.0017129344632849097, -0.12494779378175735, 0.0934644564986229, 0.014967777766287327, -0.09412139654159546, 0.05913123860955238, 0.07163394242525101, 0.03294749930500984, 0.006808740086853504, 0.04547327384352684, -0.0364118292927742, 0.030112233012914658, 0.06501468271017075, -0.0715121328830719, 0.04247138649225235, 0.07253268361091614, -0.0575287900865078, 0.13435077667236328, 0.0849154070019722, 0.038203027099370956, 0.07196131348609924, -0.10684842616319656, 0.13063564896583557, -0.02014387957751751, -0.050650689750909805, 0.059084851294755936, 0.1430433988571167, 0.0028925638180226088, 0.12119153141975403, -0.14667609333992004, 0.15125389397144318, 0.18129310011863708, -0.31137245893478394, -0.05292806774377823, 0.14816227555274963, 0.09873099625110626, 0.02743636816740036, 0.060207217931747437, 0.0749608650803566, -0.009291152469813824, 0.015535606071352959, 0.11484087258577347, -0.06311085820198059, -0.0882175862789154, 0.027218636125326157, -0.08158993721008301, -0.034129660576581955, 0.3341834247112274, 0.0007738075801171362, 0.011077087372541428, -0.0034095782320946455, -0.10655879974365234, 0.08917457610368729, -0.011537534184753895, -0.05347451940178871, 0.028600133955478668, 0.0623292475938797, 0.0015045279869809747, -0.1411283314228058, -0.04881995543837547, -0.12874853610992432, -0.15462449193000793, 0.021174535155296326, 0.01564088463783264, 0.04315921664237976, -0.16074107587337494, 0.05621093511581421, -0.029642434790730476, -0.038162339478731155, 0.010760931298136711, -0.06868353486061096, -0.09249408543109894, -0.025319524109363556, -0.10053637623786926, -0.17736941576004028, 0.08222731947898865, 0.04057405889034271, 0.10556796193122864, -0.015941627323627472, -0.04309292510151863, 0.009922911413013935, 0.02603052742779255, 0.03960378095507622, -0.11982163041830063, -0.05201924219727516, 0.0902850478887558, -0.0902024582028389, -0.10585957020521164, -0.015965349972248077, -0.150969997048378, -0.018955036997795105, 0.0019040326587855816, 0.04463041573762894, 0.05166691541671753, 0.1675269454717636, -0.0086930301040411, -0.08802442997694016, -0.0741647481918335, -0.07736483961343765, -0.06042793393135071, -0.006747439503669739, -0.10890058428049088, 0.0016295871464535594, 0.0744004175066948, 0.01591361314058304, -0.1489126980304718, -0.11269138008356094, -0.0722498670220375, -0.06376016139984131, -0.06484496593475342, -0.0974796861410141, 0.0009119802271015942, -0.10776802897453308, -0.0060530370101332664, -0.19201523065567017, -0.3714071214199066, -0.03931314870715141, 0.0682820975780487, -0.05881226435303688, -0.0617416650056839, -0.06967005133628845, -0.018371151760220528, -0.05254290997982025, -0.0420711450278759, 0.12188887596130371, -0.060492195188999176, 0.04503975063562393, -0.10438201576471329, 0.041129738092422485, -0.08865723013877869, 0.07500476390123367, -0.09997384250164032, -0.03321904316544533, 0.021262168884277344, 0.15575003623962402, 0.052938174456357956, 0.12848946452140808, -0.08721302449703217, -0.018962979316711426, -0.12163998186588287, 0.03346523270010948, 0.02471338026225567, 0.1518128216266632, -0.15512138605117798, -0.0034355190582573414, 0.16651767492294312, -0.0017009840812534094, -0.18807795643806458, 0.10849706083536148, -0.040036194026470184, 0.17604824900627136, 0.13792483508586884, 0.13802622258663177, 0.08508095145225525, -0.0654357522726059, 0.1579466313123703, 0.01688283123075962, -0.11868220567703247, -0.10349272936582565, 0.03958224505186081, 0.034916654229164124, -0.16905522346496582, 0.03237567096948624, -0.005957387387752533, 0.039804402738809586, -0.05438387393951416, -0.051647357642650604, -0.032772328704595566, -0.08326103538274765, -0.00016648051678203046, -0.033325567841529846, 0.10920291393995285, -0.0336601622402668, -0.05349021032452583, 0.017126988619565964, 0.00333221978507936, -0.025502122938632965, 0.004613265860825777, -0.09970781207084656, 0.055020444095134735, 0.0017840135842561722, 0.08870493620634079, -0.20269621908664703, -0.07311149686574936, 0.0016973091987892985, 0.18353646993637085, 0.049407072365283966, 0.08228285610675812, 0.01650042086839676, -0.05380190908908844, -0.010666726157069206, -0.0004892245051451027, 0.07458268851041794, 0.010249020531773567, -0.10635064542293549, -0.061223335564136505, 0.00005871853500138968, -0.04697785526514053, -0.08905451744794846, -0.029448162764310837, 0.03091413713991642, -0.04826575517654419, 0.10239959508180618, 0.023300442844629288, 0.05641216039657593, 0.05377217009663582, 0.02459542639553547, -0.0024522182065993547, -0.031311456114053726, 0.10461551696062088, 0.03752492740750313, -0.046162769198417664, 0.11433286219835281, -0.06433133780956268, 0.21102982759475708, 0.12419041991233826, -0.17444327473640442, -0.006151475012302399, -0.0325443260371685, -0.04602568596601486, 0.0000443191675003618, -0.04955725371837616, -0.06926814466714859, 0.14018484950065613, -0.019669322296977043, 0.1505451798439026, -0.1239149272441864, -0.0505494549870491, 0.009969521313905716, -0.017721589654684067, 0.025275159627199173, 0.03594375401735306, 0.05155956745147705, -0.223625048995018, 0.08998634666204453, 0.07832688838243484, 0.002618699334561825, 0.15135599672794342, 0.015889691188931465, -0.03434048965573311, -0.00040847083437256515, -0.03067096695303917, -0.050735265016555786, -0.038625288754701614, -0.17972400784492493, -0.045017268508672714, 0.04110569506883621, 0.04338206723332405, 0.07524392008781433, -0.025569183751940727, 0.004784321412444115, -0.004714488051831722, -0.029457898810505867, 0.0003340993425808847, 0.05919814482331276, 0.0019968899432569742, 0.13259196281433105, 0.061448365449905396, -0.024950169026851654, 0.05675138160586357, -0.036503538489341736, -0.1367182731628418, 0.18347522616386414, -0.06595179438591003, -0.29457852244377136, 0.003393077990040183, -0.057989683002233505, -0.05464375019073486, 0.05742240324616432, 0.06315188854932785, -0.08076439797878265, -0.021906830370426178, -0.09742919355630875, -0.043451957404613495, 0.02483845129609108, 0.02965758927166462, -0.09044435620307922, 0.0722133070230484, -0.015833595767617226, -0.09692145884037018, -0.029565883800387383, -0.0638708770275116, -0.09668881446123123, 0.030561372637748718, -0.20559504628181458, 0.08878922462463379, 0.10465306043624878, -0.06012309715151787, 0.09433645009994507, -0.05194919928908348, 0.16205085813999176, 0.01944883167743683, 0.00637913728132844, 0.18669526278972626, 0.04991958662867546, -0.005739824380725622, 0.07944059371948242, -0.01791341044008732, -0.09948240220546722, 0.09877140820026398, 0.04308585077524185, -0.060970790684223175, -0.2615278363227844, -0.11488006263971329, -0.07135388255119324, 0.003767447080463171, 0.10655755549669266, 0.06446993350982666, 0.08905807137489319, 0.042384736239910126, 0.012476216070353985, 0.05231470987200737, 0.04905370995402336, 0.035541046410799026, 0.12466857582330704, -0.03944191709160805, 0.1161285787820816, -0.025400007143616676, -0.040027741342782974, 0.055668942630290985, 0.06916962563991547, 0.19992513954639435, -0.001000825664959848, 0.06375962495803833, 0.03235567733645439, 0.0536658838391304, 0.04547799378633499, 0.1345280110836029, 0.029453404247760773, 0.03088989667594433, -0.04439998045563698, -0.03328587859869003, -0.046789951622486115, 0.047367263585329056, 0.01809137500822544, -0.09559539705514908, -0.10228831321001053, 0.05602032691240311, 0.08414223790168762, 0.2400396168231964, 0.010534459725022316, -0.2270810306072235, 0.0032447820995002985, 0.0020782561041414738, -0.044488511979579926, -0.0694582611322403, 0.14106886088848114, 0.011383851058781147, -0.11827436089515686, -0.003936460241675377, -0.008789281360805035, 0.20143559575080872, 0.05646128952503204, 0.03056771494448185, 0.0311838760972023, -0.004530883394181728, -0.010662449523806572, 0.10724664479494095, -0.3806115686893463, 0.19755622744560242, 0.018906166777014732, 0.04073827341198921, -0.13229599595069885, -0.04170912504196167, 0.0340241901576519, 0.0432780385017395, 0.16346780955791473, 0.009493455290794373, 0.033895425498485565, 0.03263652324676514, -0.1005336195230484, 0.06676439940929413, -0.00029342120978981256, -0.06501439213752747, 0.06208121404051781, -0.06293883919715881, -0.030924810096621513, 0.014202909544110298, 0.12671086192131042, -0.13254237174987793, -0.06251896172761917, -0.018893202766776085, 0.03007696010172367, 0.03309468924999237, -0.01751401089131832, -0.020045822486281395, 0.048522502183914185, 0.19721396267414093, 0.045788783580064774, -0.0741477832198143, -0.15248118340969086, 0.07375861704349518, 0.023011187091469765, -0.019032848998904228, 0.03767664358019829, 0.0051031955517828465, -0.0010186745785176754, 0.06024762988090515, -0.21570199728012085, 0.14051398634910583, -0.0964890569448471, -0.026591798290610313, -0.03950394317507744, 0.07933337986469269, -0.0183693990111351, 0.02978724241256714, 0.06634767353534698, -0.024175291880965233, -0.06540331244468689, -0.028011703863739967, -0.007723793387413025, 0.03221165016293526, 0.057616762816905975, 0.11637713760137558, -0.10886895656585693, -0.05705496296286583, -0.007939952425658703, 0.01421183068305254, 0.19090887904167175, -0.081430584192276, -0.023064024746418, 0.11990881711244583, 0.09158700704574585, -0.10157085210084915, -0.25910472869873047, -0.006878040265291929, -0.0036404235288500786, 0.0038429575506597757, -0.05354699864983559, -0.14648637175559998, 0.07391930371522903, 0.024463513866066933, -0.006650030612945557, -0.13824541866779327, -0.1962004452943802, -0.12192574888467789, 0.13786157965660095, 0.06970544159412384, 0.35344699025154114, -0.09963860362768173, -0.029645008966326714, -0.05317409709095955, -0.09955070912837982, 0.23699726164340973, -0.15517312288284302, 0.05222070962190628, -0.0008520019473508, -0.021173549816012383, 0.03988366201519966, -0.017839673906564713, -0.04990825057029724, 0.028092900291085243, 0.020839927718043327, -0.06820909678936005, -0.002336028264835477, 0.12444332987070084, -0.00885988399386406, 0.1494915783405304, 0.0008487062295898795, 0.13443513214588165, -0.0603620707988739, -0.07080136984586716, -0.07236853986978531, 0.03724082186818123, 0.027957962825894356, -0.07298397272825241, 0.04986530542373657, 0.012571556493639946, 0.09962693601846695, 0.006766797974705696, 0.06201605498790741, -0.0788249671459198, 0.08155225217342377, 0.15125131607055664, 0.19084739685058594, -0.11768394708633423, 0.014821846038103104, 0.013869385235011578, -0.053806863725185394, 0.07425280660390854, -0.17834383249282837, 0.041951265186071396, 0.10275881737470627, -0.011384124867618084, 0.09150869399309158, 0.08144403994083405, 0.01598592847585678, -0.009992694482207298, 0.04926472157239914, -0.15740498900413513, -0.11118149757385254, -0.12837669253349304, -0.043730951845645905, -0.02788078598678112, 0.030029091984033585, 0.1342799812555313, -0.07031305879354477, -0.01830153353512287, -0.009494048543274403, -0.03208409622311592, -0.04794413596391678, 0.13264459371566772, 0.010011664591729641, 0.00548661220818758, -0.08717211335897446, 0.11568549275398254, 0.04028208926320076, -0.04209694266319275, 0.06204689294099808, 0.10296077281236649, -0.1283416897058487, -0.10992899537086487, 0.004606857895851135, 0.15710240602493286, -0.051122330129146576, -0.05119558051228523, -0.0968574807047844, -0.08379475772380829, 0.11962065100669861, 0.14605502784252167, 0.06174363195896149, 0.06664761155843735, -0.0605585090816021, -0.035649076104164124, -0.0030485140159726143, 0.10053864866495132, 0.049986422061920166, -0.028461040928959846, -0.10298031568527222, 0.04793676361441612, -0.005563674494624138, 0.11872929334640503, -0.09751354157924652, -0.04358326271176338, -0.11636678874492645, 0.051670901477336884, -0.18654164671897888, -0.025050649419426918, -0.09060357511043549, 0.00023375946329906583, -0.05468641594052315, -0.022427916526794434, -0.06776372343301773, 0.039591897279024124, -0.06372570991516113, 0.003695577848702669, 0.007594872731715441, 0.06061221659183502, -0.03312499821186066, -0.03622165322303772, -0.01453554630279541, -0.013161752372980118, 0.1189628392457962, 0.10756859928369522, -0.13064563274383545, 0.028492320328950882, -0.1101660281419754, -0.033476125448942184, 0.023578858003020287, 0.0730619877576828, 0.03709006682038307, 0.0056908042170107365, 0.04245435819029808, 0.08798976987600327, 0.0032347713131457567, 0.008251786231994629, 0.060150545090436935, -0.07843322306871414, -0.01904105953872204, -0.039244502782821655, -0.02896038442850113, -0.06378979980945587, -0.02443031780421734, 0.07188011705875397, 0.10120058804750443, 0.0888434574007988, -0.046730685979127884, 0.04530879110097885, -0.1321856677532196, 0.01648041047155857, 0.010114041157066822, -0.09666027128696442, -0.14170171320438385, -0.0728374719619751, 0.06222446635365486, -0.032072801142930984, 0.21843987703323364, 0.10028482973575592, 0.08588670194149017, 0.026794854551553726, 0.12262635678052902, 0.13682037591934204, 0.028858831152319908, 0.15925246477127075, -0.0039056914392858744, -0.03759155049920082, -0.006454679649323225, 0.07537528872489929, 0.02004428394138813, 0.10966537147760391, 0.1103924810886383, 0.06846192479133606, 0.09284596145153046, 0.051480911672115326, 0.040691956877708435, 0.043421465903520584, -0.048595838248729706, -0.04143897816538811, -0.008195987902581692, 0.07680226117372513, -0.07515399158000946, -0.07015363872051239, 0.14670173823833466, -0.06356364488601685, 0.039187874644994736, 0.0032016620971262455, -0.023885097354650497, -0.12871825695037842, -0.2169605940580368, -0.11109609156847, -0.12449198216199875, -0.04429619759321213, -0.16815033555030823, -0.016044791787862778, -0.015458434820175171, 0.02580721490085125, -0.03197168931365013, 0.08031291514635086, -0.0513436421751976, -0.08022155612707138, 0.09766086935997009, -0.06816882640123367, 0.053412504494190216, -0.05675780773162842, -0.006436794530600309, 0.008212004788219929, 0.001708818250335753, -0.0502929724752903, 0.06155063584446907, -0.026821281760931015, 0.04074952006340027, -0.0789605975151062, -0.050365932285785675, -0.048448171466588974, 0.019155921414494514, -0.045548051595687866, 0.05766923353075981, 0.05587313324213028, -0.10920967161655426, 0.052133165299892426, 0.24796870350837708, -0.03879782557487488, -0.03642837703227997, -0.17920319736003876, 0.14137135446071625, 0.06285294145345688, 0.01990879885852337, 0.05674859881401062, 0.005587015300989151, -0.04000172019004822, 0.3214752972126007, 0.2232431173324585, -0.0940014123916626, -0.001401323126628995, 0.032077059149742126, -0.0029140186961740255, 0.04978270083665848, 0.1456911265850067, 0.05564079061150551, 0.21695257723331451, -0.02343052066862583, -0.008549569174647331, 0.005132440943270922, -0.0604737289249897, -0.03798462823033333, 0.023236146196722984, 0.09155768156051636, -0.0667944923043251, 0.007818769663572311, 0.10838977247476578, -0.20991641283035278, 0.08792625367641449, -0.06710314750671387, -0.0954042300581932, -0.08775869756937027, -0.09129097312688828, -0.061213746666908264, 0.05177313834428787, 0.07833156734704971, -0.05678052082657814, 0.04240644350647926, -0.01248181238770485, 0.005435170605778694, -0.18059654533863068, -0.09907633066177368, 0.10086842626333237, 0.07562318444252014, 0.09460794925689697, -0.01361129991710186, 0.0465344600379467, 0.061317235231399536, 0.013600200414657593, -0.1216587945818901, 0.05813458189368248, -0.024338945746421814, 0.001456014346331358, 0.10642140358686447, -0.062257517129182816, -0.029824716970324516, 0.03045003116130829, 0.03735160827636719, -0.10519750416278839, 0.025238044559955597, 0.026334309950470924, 0.009133719839155674, -0.09221337735652924, 0.07049732655286789, -0.0749671459197998, 0.080876924097538, 0.12738698720932007, -0.041905421763658524, 0.03221181407570839, -0.08910879492759705, 0.030222728848457336, 0.019200339913368225, -0.10525348782539368, -0.03967959061264992, -0.1568468064069748, -0.06707330048084259, 0.02184991165995598, -0.06456150859594345, -0.1485566347837448, -0.004499278496950865, -0.09853282570838928, -0.015628185123205185, -0.08386103063821793, 0.1112486720085144, 0.1288033127784729, 0.018936283886432648, -0.0088672935962677, -0.01518254540860653, 0.00945657305419445, 0.11685440689325333, -0.16846786439418793, -0.1284887194633484 ]
null
null
transformers
# Query Generation This model is the t5-base model from [docTTTTTquery](https://github.com/castorini/docTTTTTquery). The T5-base model was trained on the [MS MARCO Passage Dataset](https://github.com/microsoft/MSMARCO-Passage-Ranking), which consists of about 500k real search queries from Bing together with the relevant passage. The model can be used for query generation to learn semantic search models without requiring annotated training data: [Synthetic Query Generation](https://github.com/UKPLab/sentence-transformers/tree/master/examples/unsupervised_learning/query_generation). ## Usage ```python from transformers import T5Tokenizer, T5ForConditionalGeneration tokenizer = T5Tokenizer.from_pretrained('model-name') model = T5ForConditionalGeneration.from_pretrained('model-name') para = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects." input_ids = tokenizer.encode(para, return_tensors='pt') outputs = model.generate( input_ids=input_ids, max_length=64, do_sample=True, top_p=0.95, num_return_sequences=3) print("Paragraph:") print(para) print("\nGenerated Queries:") for i in range(len(outputs)): query = tokenizer.decode(outputs[i], skip_special_tokens=True) print(f'{i + 1}: {query}') ```
{}
text2text-generation
BeIR/query-gen-msmarco-t5-large-v1
[ "transformers", "pytorch", "jax", "t5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #jax #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Query Generation This model is the t5-base model from docTTTTTquery. The T5-base model was trained on the MS MARCO Passage Dataset, which consists of about 500k real search queries from Bing together with the relevant passage. The model can be used for query generation to learn semantic search models without requiring annotated training data: Synthetic Query Generation. ## Usage
[ "# Query Generation\nThis model is the t5-base model from docTTTTTquery.\n\nThe T5-base model was trained on the MS MARCO Passage Dataset, which consists of about 500k real search queries from Bing together with the relevant passage.\n\nThe model can be used for query generation to learn semantic search models without requiring annotated training data: Synthetic Query Generation.", "## Usage" ]
[ "TAGS\n#transformers #pytorch #jax #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Query Generation\nThis model is the t5-base model from docTTTTTquery.\n\nThe T5-base model was trained on the MS MARCO Passage Dataset, which consists of about 500k real search queries from Bing together with the relevant passage.\n\nThe model can be used for query generation to learn semantic search models without requiring annotated training data: Synthetic Query Generation.", "## Usage" ]
[ 51, 87, 3 ]
[ "passage: TAGS\n#transformers #pytorch #jax #t5 #text2text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Query Generation\nThis model is the t5-base model from docTTTTTquery.\n\nThe T5-base model was trained on the MS MARCO Passage Dataset, which consists of about 500k real search queries from Bing together with the relevant passage.\n\nThe model can be used for query generation to learn semantic search models without requiring annotated training data: Synthetic Query Generation.## Usage" ]
[ 0.0035321186296641827, 0.07132449746131897, -0.0034816828556358814, 0.05618482828140259, 0.16366024315357208, 0.0070151109248399734, 0.05777270719408989, 0.09940236061811447, 0.03087661601603031, -0.0799681693315506, 0.09968629479408264, 0.09432600438594818, -0.02423050068318844, 0.1521737277507782, -0.04812584072351456, -0.24514101445674896, 0.00973388273268938, 0.02144435979425907, -0.05103942006826401, 0.11938455700874329, 0.09747091680765152, -0.05866660177707672, 0.13916876912117004, -0.0017129344632849097, -0.12494779378175735, 0.0934644564986229, 0.014967777766287327, -0.09412139654159546, 0.05913123860955238, 0.07163394242525101, 0.03294749930500984, 0.006808740086853504, 0.04547327384352684, -0.0364118292927742, 0.030112233012914658, 0.06501468271017075, -0.0715121328830719, 0.04247138649225235, 0.07253268361091614, -0.0575287900865078, 0.13435077667236328, 0.0849154070019722, 0.038203027099370956, 0.07196131348609924, -0.10684842616319656, 0.13063564896583557, -0.02014387957751751, -0.050650689750909805, 0.059084851294755936, 0.1430433988571167, 0.0028925638180226088, 0.12119153141975403, -0.14667609333992004, 0.15125389397144318, 0.18129310011863708, -0.31137245893478394, -0.05292806774377823, 0.14816227555274963, 0.09873099625110626, 0.02743636816740036, 0.060207217931747437, 0.0749608650803566, -0.009291152469813824, 0.015535606071352959, 0.11484087258577347, -0.06311085820198059, -0.0882175862789154, 0.027218636125326157, -0.08158993721008301, -0.034129660576581955, 0.3341834247112274, 0.0007738075801171362, 0.011077087372541428, -0.0034095782320946455, -0.10655879974365234, 0.08917457610368729, -0.011537534184753895, -0.05347451940178871, 0.028600133955478668, 0.0623292475938797, 0.0015045279869809747, -0.1411283314228058, -0.04881995543837547, -0.12874853610992432, -0.15462449193000793, 0.021174535155296326, 0.01564088463783264, 0.04315921664237976, -0.16074107587337494, 0.05621093511581421, -0.029642434790730476, -0.038162339478731155, 0.010760931298136711, -0.06868353486061096, -0.09249408543109894, -0.025319524109363556, -0.10053637623786926, -0.17736941576004028, 0.08222731947898865, 0.04057405889034271, 0.10556796193122864, -0.015941627323627472, -0.04309292510151863, 0.009922911413013935, 0.02603052742779255, 0.03960378095507622, -0.11982163041830063, -0.05201924219727516, 0.0902850478887558, -0.0902024582028389, -0.10585957020521164, -0.015965349972248077, -0.150969997048378, -0.018955036997795105, 0.0019040326587855816, 0.04463041573762894, 0.05166691541671753, 0.1675269454717636, -0.0086930301040411, -0.08802442997694016, -0.0741647481918335, -0.07736483961343765, -0.06042793393135071, -0.006747439503669739, -0.10890058428049088, 0.0016295871464535594, 0.0744004175066948, 0.01591361314058304, -0.1489126980304718, -0.11269138008356094, -0.0722498670220375, -0.06376016139984131, -0.06484496593475342, -0.0974796861410141, 0.0009119802271015942, -0.10776802897453308, -0.0060530370101332664, -0.19201523065567017, -0.3714071214199066, -0.03931314870715141, 0.0682820975780487, -0.05881226435303688, -0.0617416650056839, -0.06967005133628845, -0.018371151760220528, -0.05254290997982025, -0.0420711450278759, 0.12188887596130371, -0.060492195188999176, 0.04503975063562393, -0.10438201576471329, 0.041129738092422485, -0.08865723013877869, 0.07500476390123367, -0.09997384250164032, -0.03321904316544533, 0.021262168884277344, 0.15575003623962402, 0.052938174456357956, 0.12848946452140808, -0.08721302449703217, -0.018962979316711426, -0.12163998186588287, 0.03346523270010948, 0.02471338026225567, 0.1518128216266632, -0.15512138605117798, -0.0034355190582573414, 0.16651767492294312, -0.0017009840812534094, -0.18807795643806458, 0.10849706083536148, -0.040036194026470184, 0.17604824900627136, 0.13792483508586884, 0.13802622258663177, 0.08508095145225525, -0.0654357522726059, 0.1579466313123703, 0.01688283123075962, -0.11868220567703247, -0.10349272936582565, 0.03958224505186081, 0.034916654229164124, -0.16905522346496582, 0.03237567096948624, -0.005957387387752533, 0.039804402738809586, -0.05438387393951416, -0.051647357642650604, -0.032772328704595566, -0.08326103538274765, -0.00016648051678203046, -0.033325567841529846, 0.10920291393995285, -0.0336601622402668, -0.05349021032452583, 0.017126988619565964, 0.00333221978507936, -0.025502122938632965, 0.004613265860825777, -0.09970781207084656, 0.055020444095134735, 0.0017840135842561722, 0.08870493620634079, -0.20269621908664703, -0.07311149686574936, 0.0016973091987892985, 0.18353646993637085, 0.049407072365283966, 0.08228285610675812, 0.01650042086839676, -0.05380190908908844, -0.010666726157069206, -0.0004892245051451027, 0.07458268851041794, 0.010249020531773567, -0.10635064542293549, -0.061223335564136505, 0.00005871853500138968, -0.04697785526514053, -0.08905451744794846, -0.029448162764310837, 0.03091413713991642, -0.04826575517654419, 0.10239959508180618, 0.023300442844629288, 0.05641216039657593, 0.05377217009663582, 0.02459542639553547, -0.0024522182065993547, -0.031311456114053726, 0.10461551696062088, 0.03752492740750313, -0.046162769198417664, 0.11433286219835281, -0.06433133780956268, 0.21102982759475708, 0.12419041991233826, -0.17444327473640442, -0.006151475012302399, -0.0325443260371685, -0.04602568596601486, 0.0000443191675003618, -0.04955725371837616, -0.06926814466714859, 0.14018484950065613, -0.019669322296977043, 0.1505451798439026, -0.1239149272441864, -0.0505494549870491, 0.009969521313905716, -0.017721589654684067, 0.025275159627199173, 0.03594375401735306, 0.05155956745147705, -0.223625048995018, 0.08998634666204453, 0.07832688838243484, 0.002618699334561825, 0.15135599672794342, 0.015889691188931465, -0.03434048965573311, -0.00040847083437256515, -0.03067096695303917, -0.050735265016555786, -0.038625288754701614, -0.17972400784492493, -0.045017268508672714, 0.04110569506883621, 0.04338206723332405, 0.07524392008781433, -0.025569183751940727, 0.004784321412444115, -0.004714488051831722, -0.029457898810505867, 0.0003340993425808847, 0.05919814482331276, 0.0019968899432569742, 0.13259196281433105, 0.061448365449905396, -0.024950169026851654, 0.05675138160586357, -0.036503538489341736, -0.1367182731628418, 0.18347522616386414, -0.06595179438591003, -0.29457852244377136, 0.003393077990040183, -0.057989683002233505, -0.05464375019073486, 0.05742240324616432, 0.06315188854932785, -0.08076439797878265, -0.021906830370426178, -0.09742919355630875, -0.043451957404613495, 0.02483845129609108, 0.02965758927166462, -0.09044435620307922, 0.0722133070230484, -0.015833595767617226, -0.09692145884037018, -0.029565883800387383, -0.0638708770275116, -0.09668881446123123, 0.030561372637748718, -0.20559504628181458, 0.08878922462463379, 0.10465306043624878, -0.06012309715151787, 0.09433645009994507, -0.05194919928908348, 0.16205085813999176, 0.01944883167743683, 0.00637913728132844, 0.18669526278972626, 0.04991958662867546, -0.005739824380725622, 0.07944059371948242, -0.01791341044008732, -0.09948240220546722, 0.09877140820026398, 0.04308585077524185, -0.060970790684223175, -0.2615278363227844, -0.11488006263971329, -0.07135388255119324, 0.003767447080463171, 0.10655755549669266, 0.06446993350982666, 0.08905807137489319, 0.042384736239910126, 0.012476216070353985, 0.05231470987200737, 0.04905370995402336, 0.035541046410799026, 0.12466857582330704, -0.03944191709160805, 0.1161285787820816, -0.025400007143616676, -0.040027741342782974, 0.055668942630290985, 0.06916962563991547, 0.19992513954639435, -0.001000825664959848, 0.06375962495803833, 0.03235567733645439, 0.0536658838391304, 0.04547799378633499, 0.1345280110836029, 0.029453404247760773, 0.03088989667594433, -0.04439998045563698, -0.03328587859869003, -0.046789951622486115, 0.047367263585329056, 0.01809137500822544, -0.09559539705514908, -0.10228831321001053, 0.05602032691240311, 0.08414223790168762, 0.2400396168231964, 0.010534459725022316, -0.2270810306072235, 0.0032447820995002985, 0.0020782561041414738, -0.044488511979579926, -0.0694582611322403, 0.14106886088848114, 0.011383851058781147, -0.11827436089515686, -0.003936460241675377, -0.008789281360805035, 0.20143559575080872, 0.05646128952503204, 0.03056771494448185, 0.0311838760972023, -0.004530883394181728, -0.010662449523806572, 0.10724664479494095, -0.3806115686893463, 0.19755622744560242, 0.018906166777014732, 0.04073827341198921, -0.13229599595069885, -0.04170912504196167, 0.0340241901576519, 0.0432780385017395, 0.16346780955791473, 0.009493455290794373, 0.033895425498485565, 0.03263652324676514, -0.1005336195230484, 0.06676439940929413, -0.00029342120978981256, -0.06501439213752747, 0.06208121404051781, -0.06293883919715881, -0.030924810096621513, 0.014202909544110298, 0.12671086192131042, -0.13254237174987793, -0.06251896172761917, -0.018893202766776085, 0.03007696010172367, 0.03309468924999237, -0.01751401089131832, -0.020045822486281395, 0.048522502183914185, 0.19721396267414093, 0.045788783580064774, -0.0741477832198143, -0.15248118340969086, 0.07375861704349518, 0.023011187091469765, -0.019032848998904228, 0.03767664358019829, 0.0051031955517828465, -0.0010186745785176754, 0.06024762988090515, -0.21570199728012085, 0.14051398634910583, -0.0964890569448471, -0.026591798290610313, -0.03950394317507744, 0.07933337986469269, -0.0183693990111351, 0.02978724241256714, 0.06634767353534698, -0.024175291880965233, -0.06540331244468689, -0.028011703863739967, -0.007723793387413025, 0.03221165016293526, 0.057616762816905975, 0.11637713760137558, -0.10886895656585693, -0.05705496296286583, -0.007939952425658703, 0.01421183068305254, 0.19090887904167175, -0.081430584192276, -0.023064024746418, 0.11990881711244583, 0.09158700704574585, -0.10157085210084915, -0.25910472869873047, -0.006878040265291929, -0.0036404235288500786, 0.0038429575506597757, -0.05354699864983559, -0.14648637175559998, 0.07391930371522903, 0.024463513866066933, -0.006650030612945557, -0.13824541866779327, -0.1962004452943802, -0.12192574888467789, 0.13786157965660095, 0.06970544159412384, 0.35344699025154114, -0.09963860362768173, -0.029645008966326714, -0.05317409709095955, -0.09955070912837982, 0.23699726164340973, -0.15517312288284302, 0.05222070962190628, -0.0008520019473508, -0.021173549816012383, 0.03988366201519966, -0.017839673906564713, -0.04990825057029724, 0.028092900291085243, 0.020839927718043327, -0.06820909678936005, -0.002336028264835477, 0.12444332987070084, -0.00885988399386406, 0.1494915783405304, 0.0008487062295898795, 0.13443513214588165, -0.0603620707988739, -0.07080136984586716, -0.07236853986978531, 0.03724082186818123, 0.027957962825894356, -0.07298397272825241, 0.04986530542373657, 0.012571556493639946, 0.09962693601846695, 0.006766797974705696, 0.06201605498790741, -0.0788249671459198, 0.08155225217342377, 0.15125131607055664, 0.19084739685058594, -0.11768394708633423, 0.014821846038103104, 0.013869385235011578, -0.053806863725185394, 0.07425280660390854, -0.17834383249282837, 0.041951265186071396, 0.10275881737470627, -0.011384124867618084, 0.09150869399309158, 0.08144403994083405, 0.01598592847585678, -0.009992694482207298, 0.04926472157239914, -0.15740498900413513, -0.11118149757385254, -0.12837669253349304, -0.043730951845645905, -0.02788078598678112, 0.030029091984033585, 0.1342799812555313, -0.07031305879354477, -0.01830153353512287, -0.009494048543274403, -0.03208409622311592, -0.04794413596391678, 0.13264459371566772, 0.010011664591729641, 0.00548661220818758, -0.08717211335897446, 0.11568549275398254, 0.04028208926320076, -0.04209694266319275, 0.06204689294099808, 0.10296077281236649, -0.1283416897058487, -0.10992899537086487, 0.004606857895851135, 0.15710240602493286, -0.051122330129146576, -0.05119558051228523, -0.0968574807047844, -0.08379475772380829, 0.11962065100669861, 0.14605502784252167, 0.06174363195896149, 0.06664761155843735, -0.0605585090816021, -0.035649076104164124, -0.0030485140159726143, 0.10053864866495132, 0.049986422061920166, -0.028461040928959846, -0.10298031568527222, 0.04793676361441612, -0.005563674494624138, 0.11872929334640503, -0.09751354157924652, -0.04358326271176338, -0.11636678874492645, 0.051670901477336884, -0.18654164671897888, -0.025050649419426918, -0.09060357511043549, 0.00023375946329906583, -0.05468641594052315, -0.022427916526794434, -0.06776372343301773, 0.039591897279024124, -0.06372570991516113, 0.003695577848702669, 0.007594872731715441, 0.06061221659183502, -0.03312499821186066, -0.03622165322303772, -0.01453554630279541, -0.013161752372980118, 0.1189628392457962, 0.10756859928369522, -0.13064563274383545, 0.028492320328950882, -0.1101660281419754, -0.033476125448942184, 0.023578858003020287, 0.0730619877576828, 0.03709006682038307, 0.0056908042170107365, 0.04245435819029808, 0.08798976987600327, 0.0032347713131457567, 0.008251786231994629, 0.060150545090436935, -0.07843322306871414, -0.01904105953872204, -0.039244502782821655, -0.02896038442850113, -0.06378979980945587, -0.02443031780421734, 0.07188011705875397, 0.10120058804750443, 0.0888434574007988, -0.046730685979127884, 0.04530879110097885, -0.1321856677532196, 0.01648041047155857, 0.010114041157066822, -0.09666027128696442, -0.14170171320438385, -0.0728374719619751, 0.06222446635365486, -0.032072801142930984, 0.21843987703323364, 0.10028482973575592, 0.08588670194149017, 0.026794854551553726, 0.12262635678052902, 0.13682037591934204, 0.028858831152319908, 0.15925246477127075, -0.0039056914392858744, -0.03759155049920082, -0.006454679649323225, 0.07537528872489929, 0.02004428394138813, 0.10966537147760391, 0.1103924810886383, 0.06846192479133606, 0.09284596145153046, 0.051480911672115326, 0.040691956877708435, 0.043421465903520584, -0.048595838248729706, -0.04143897816538811, -0.008195987902581692, 0.07680226117372513, -0.07515399158000946, -0.07015363872051239, 0.14670173823833466, -0.06356364488601685, 0.039187874644994736, 0.0032016620971262455, -0.023885097354650497, -0.12871825695037842, -0.2169605940580368, -0.11109609156847, -0.12449198216199875, -0.04429619759321213, -0.16815033555030823, -0.016044791787862778, -0.015458434820175171, 0.02580721490085125, -0.03197168931365013, 0.08031291514635086, -0.0513436421751976, -0.08022155612707138, 0.09766086935997009, -0.06816882640123367, 0.053412504494190216, -0.05675780773162842, -0.006436794530600309, 0.008212004788219929, 0.001708818250335753, -0.0502929724752903, 0.06155063584446907, -0.026821281760931015, 0.04074952006340027, -0.0789605975151062, -0.050365932285785675, -0.048448171466588974, 0.019155921414494514, -0.045548051595687866, 0.05766923353075981, 0.05587313324213028, -0.10920967161655426, 0.052133165299892426, 0.24796870350837708, -0.03879782557487488, -0.03642837703227997, -0.17920319736003876, 0.14137135446071625, 0.06285294145345688, 0.01990879885852337, 0.05674859881401062, 0.005587015300989151, -0.04000172019004822, 0.3214752972126007, 0.2232431173324585, -0.0940014123916626, -0.001401323126628995, 0.032077059149742126, -0.0029140186961740255, 0.04978270083665848, 0.1456911265850067, 0.05564079061150551, 0.21695257723331451, -0.02343052066862583, -0.008549569174647331, 0.005132440943270922, -0.0604737289249897, -0.03798462823033333, 0.023236146196722984, 0.09155768156051636, -0.0667944923043251, 0.007818769663572311, 0.10838977247476578, -0.20991641283035278, 0.08792625367641449, -0.06710314750671387, -0.0954042300581932, -0.08775869756937027, -0.09129097312688828, -0.061213746666908264, 0.05177313834428787, 0.07833156734704971, -0.05678052082657814, 0.04240644350647926, -0.01248181238770485, 0.005435170605778694, -0.18059654533863068, -0.09907633066177368, 0.10086842626333237, 0.07562318444252014, 0.09460794925689697, -0.01361129991710186, 0.0465344600379467, 0.061317235231399536, 0.013600200414657593, -0.1216587945818901, 0.05813458189368248, -0.024338945746421814, 0.001456014346331358, 0.10642140358686447, -0.062257517129182816, -0.029824716970324516, 0.03045003116130829, 0.03735160827636719, -0.10519750416278839, 0.025238044559955597, 0.026334309950470924, 0.009133719839155674, -0.09221337735652924, 0.07049732655286789, -0.0749671459197998, 0.080876924097538, 0.12738698720932007, -0.041905421763658524, 0.03221181407570839, -0.08910879492759705, 0.030222728848457336, 0.019200339913368225, -0.10525348782539368, -0.03967959061264992, -0.1568468064069748, -0.06707330048084259, 0.02184991165995598, -0.06456150859594345, -0.1485566347837448, -0.004499278496950865, -0.09853282570838928, -0.015628185123205185, -0.08386103063821793, 0.1112486720085144, 0.1288033127784729, 0.018936283886432648, -0.0088672935962677, -0.01518254540860653, 0.00945657305419445, 0.11685440689325333, -0.16846786439418793, -0.1284887194633484 ]
null
null
transformers
# SPARTA Re-Implementation of [SPARTA: Efficient Open-Domain Question Answering via Sparse Transformer Matching Retrieval](https://arxiv.org/abs/2009.13013). It is the re-implementation we used for [BEIR: A Heterogenous Benchmark for Zero-shot Evaluation of Information Retrieval Models](https://arxiv.org/abs/2104.08663). Also have a look at our BEIR repository: https://github.com/UKPLab/beir Have a look at https://github.com/nreimers/beir-sparta for the training and inference code of this SPARTA model
{}
feature-extraction
BeIR/sparta-msmarco-distilbert-base-v1
[ "transformers", "pytorch", "distilbert", "feature-extraction", "arxiv:2009.13013", "arxiv:2104.08663", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[ "2009.13013", "2104.08663" ]
[]
TAGS #transformers #pytorch #distilbert #feature-extraction #arxiv-2009.13013 #arxiv-2104.08663 #endpoints_compatible #region-us
# SPARTA Re-Implementation of SPARTA: Efficient Open-Domain Question Answering via Sparse Transformer Matching Retrieval. It is the re-implementation we used for BEIR: A Heterogenous Benchmark for Zero-shot Evaluation of Information Retrieval Models. Also have a look at our BEIR repository: URL Have a look at URL for the training and inference code of this SPARTA model
[ "# SPARTA\nRe-Implementation of SPARTA: Efficient Open-Domain Question Answering via Sparse Transformer Matching Retrieval. It is the re-implementation we used for BEIR: A Heterogenous Benchmark for Zero-shot Evaluation of Information Retrieval Models.\n\nAlso have a look at our BEIR repository: URL\n\n\nHave a look at URL for the training and inference code of this SPARTA model" ]
[ "TAGS\n#transformers #pytorch #distilbert #feature-extraction #arxiv-2009.13013 #arxiv-2104.08663 #endpoints_compatible #region-us \n", "# SPARTA\nRe-Implementation of SPARTA: Efficient Open-Domain Question Answering via Sparse Transformer Matching Retrieval. It is the re-implementation we used for BEIR: A Heterogenous Benchmark for Zero-shot Evaluation of Information Retrieval Models.\n\nAlso have a look at our BEIR repository: URL\n\n\nHave a look at URL for the training and inference code of this SPARTA model" ]
[ 48, 98 ]
[ "passage: TAGS\n#transformers #pytorch #distilbert #feature-extraction #arxiv-2009.13013 #arxiv-2104.08663 #endpoints_compatible #region-us \n# SPARTA\nRe-Implementation of SPARTA: Efficient Open-Domain Question Answering via Sparse Transformer Matching Retrieval. It is the re-implementation we used for BEIR: A Heterogenous Benchmark for Zero-shot Evaluation of Information Retrieval Models.\n\nAlso have a look at our BEIR repository: URL\n\n\nHave a look at URL for the training and inference code of this SPARTA model" ]
[ -0.07133377343416214, 0.03410220518708229, -0.0038397801108658314, 0.07805600762367249, 0.13684618473052979, -0.011236834339797497, 0.052567340433597565, 0.07208997756242752, -0.04085112735629082, 0.004803827963769436, 0.08282213658094406, 0.10588844120502472, -0.017630266025662422, 0.11913774907588959, -0.08473041653633118, -0.14565540850162506, 0.03356939181685448, 0.04261048138141632, -0.041409339755773544, 0.121999591588974, 0.1130940392613411, -0.06048363819718361, 0.08095236122608185, 0.03217972815036774, -0.16236701607704163, 0.08223140984773636, -0.015714148059487343, -0.05290645360946655, 0.08706489205360413, -0.02932921051979065, 0.12495870143175125, 0.03793032839894295, 0.06888105720281601, -0.054728589951992035, 0.028823690488934517, -0.009673958644270897, -0.00665227510035038, 0.06131851673126221, -0.05923834443092346, -0.08203355967998505, 0.03786754608154297, 0.061057575047016144, 0.015209733508527279, 0.013994282111525536, -0.1260731816291809, 0.007405685726553202, -0.009545453824102879, 0.04825848340988159, 0.020598312839865685, 0.12312914431095123, 0.016239605844020844, 0.1847393959760666, -0.1212458685040474, 0.08243850618600845, 0.1401343196630478, -0.2153279185295105, -0.012148115783929825, 0.010603758506476879, -0.016510188579559326, 0.005251436028629541, -0.02928720787167549, 0.033806849271059036, 0.04151957109570503, -0.0005233780830167234, 0.01938570849597454, -0.0681651309132576, -0.12480820715427399, 0.03999675437808037, -0.13947662711143494, -0.03706534206867218, 0.3505447506904602, 0.03804082050919533, -0.03441440314054489, 0.08359984308481216, -0.11104442924261093, 0.09709581732749939, -0.028877070173621178, -0.028313232585787773, -0.004779435694217682, 0.015680532902479172, -0.06528142094612122, -0.053017228841781616, -0.0649154782295227, -0.14542295038700104, -0.11718539148569107, 0.15749335289001465, 0.033107008785009384, 0.07569172233343124, -0.13050591945648193, 0.06267458200454712, 0.046624258160591125, -0.07605291157960892, -0.010430257767438889, -0.08239948749542236, -0.04153289645910263, 0.001738098100759089, -0.07618943601846695, 0.0060258666053414345, 0.06651511788368225, 0.30334436893463135, 0.13247346878051758, -0.00104193773586303, -0.01101731602102518, 0.05118459090590477, 0.02674424834549427, 0.08478105813264847, -0.11419612914323807, -0.01625889539718628, 0.04955611750483513, -0.055775489658117294, -0.018941964954137802, 0.005174423102289438, -0.052253857254981995, -0.035361967980861664, 0.018617931753396988, 0.054994259029626846, 0.037146925926208496, 0.12073025852441788, -0.05972232297062874, -0.04648340865969658, -0.02227860689163208, -0.10745717585086823, -0.04547755420207977, 0.02091553620994091, -0.09112419933080673, 0.03697481378912926, 0.09004943072795868, -0.04187244176864624, -0.0777725949883461, -0.05458293855190277, -0.09761383384466171, -0.0023961644619703293, -0.043957095593214035, -0.11647924780845642, -0.01355221588164568, -0.13807815313339233, 0.047523174434900284, -0.17322179675102234, -0.2446596771478653, -0.02932080812752247, 0.03218725696206093, -0.03932173177599907, 0.02537023462355137, -0.08455725014209747, -0.053520627319812775, -0.04799915477633476, -0.043235234916210175, -0.0653020590543747, -0.05070554465055466, 0.036114372313022614, 0.018163461238145828, 0.07507423311471939, -0.0950518324971199, 0.04567396268248558, -0.11790723353624344, -0.024629026651382446, -0.026144076138734818, 0.07503242790699005, -0.08539292216300964, 0.07781784236431122, -0.10647325962781906, -0.05347432941198349, -0.07569553703069687, -0.0013463632203638554, 0.08002044260501862, 0.19954918324947357, -0.08426924794912338, -0.028171729296445847, 0.08491984754800797, -0.14776109158992767, -0.1111888736486435, 0.06577420234680176, -0.06973834335803986, 0.10946909338235855, 0.05587245896458626, 0.13730156421661377, 0.04700804501771927, -0.05884217470884323, 0.08108651638031006, 0.023099154233932495, -0.19737601280212402, -0.06596854329109192, 0.04483752325177193, 0.06656135618686676, -0.04166173189878464, 0.041485391557216644, -0.11118791997432709, 0.059154216200113297, -0.07920344918966293, -0.05458879470825195, -0.01979314535856247, -0.041802484542131424, -0.05315056070685387, 0.01931372843682766, 0.08876483887434006, 0.013287726789712906, 0.021450882777571678, 0.16544081270694733, 0.06875676661729813, -0.04609013721346855, 0.022314587607979774, -0.12048359960317612, 0.09168215095996857, -0.1312340945005417, -0.001718129264190793, -0.23292025923728943, 0.04802237078547478, -0.0558287613093853, 0.14398950338363647, 0.04167373478412628, 0.10082530975341797, 0.05322030559182167, -0.09393353015184402, 0.0063959225080907345, -0.028082603588700294, 0.013182135298848152, 0.05484927445650101, 0.0036443695425987244, -0.11330763250589371, 0.002246514894068241, -0.028412800282239914, 0.044759467244148254, 0.01676364243030548, -0.01660129427909851, -0.0564226359128952, 0.10028409212827682, -0.020659189671278, 0.02893422730267048, 0.012774134054780006, 0.003851602552458644, -0.03698553144931793, -0.018539611250162125, 0.09102745354175568, 0.05154554918408394, -0.0758076086640358, 0.03104107268154621, 0.014284523203969002, 0.1376127153635025, 0.10458018630743027, -0.21203778684139252, -0.0652649849653244, 0.03511243686079979, -0.04801816865801811, 0.0055193351581692696, 0.017341919243335724, -0.03369494527578354, 0.12722094357013702, -0.033209241926670074, 0.10815735161304474, -0.04139268770813942, 0.010099229402840137, -0.01988093927502632, -0.05753997340798378, 0.026706743985414505, 0.11817748099565506, 0.09060415625572205, -0.27217501401901245, 0.1104857474565506, 0.018476011231541634, -0.06377892196178436, 0.02297034300863743, -0.0736447125673294, -0.060178425163030624, -0.015375182032585144, 0.024814747273921967, -0.002496313536539674, 0.08503882586956024, -0.0875120609998703, -0.04741818830370903, 0.04991495609283447, -0.0015706128906458616, 0.042509038001298904, -0.05982990935444832, 0.026361534371972084, 0.023953355848789215, -0.001108496799133718, -0.11736587435007095, -0.01195843517780304, -0.012973436154425144, 0.07693150639533997, 0.026794055476784706, -0.11922190338373184, 0.01039358600974083, -0.009675832465291023, -0.07376611232757568, 0.1717245876789093, -0.045349083840847015, -0.11807483434677124, -0.12553824484348297, 0.05713307112455368, -0.021383268758654594, 0.0229722261428833, 0.017781339585781097, -0.01268346980214119, -0.03472140431404114, -0.02596256695687771, 0.01944603957235813, -0.08162084221839905, -0.01656232960522175, 0.02395871840417385, 0.044703081250190735, 0.045375462621450424, -0.12342314422130585, -0.023785671219229698, -0.11236497014760971, -0.013515396043658257, 0.06154441833496094, -0.13741451501846313, 0.12256856262683868, 0.019220061600208282, -0.00842313189059496, 0.05138351023197174, -0.00973021425306797, 0.14552105963230133, 0.0036932167131453753, -0.001136949984356761, 0.12355201691389084, -0.07194273918867111, 0.005551917478442192, 0.10589811205863953, -0.015610164031386375, -0.07987632602453232, 0.050796519964933395, 0.0004596714279614389, -0.040032219141721725, -0.1636246293783188, -0.09991360455751419, -0.08696945011615753, 0.00431883754208684, 0.052592359483242035, 0.01299246121197939, -0.08053673058748245, 0.09197747707366943, 0.0253800917416811, -0.030122073367238045, -0.0039557586424052715, 0.07277946174144745, 0.15866826474666595, -0.01679021120071411, 0.12426286190748215, -0.05568413436412811, -0.1275678277015686, 0.03920099139213562, -0.03421531990170479, 0.24729016423225403, -0.0004337222489994019, 0.07228198647499084, 0.09866620600223541, 0.06355149298906326, 0.06748979538679123, 0.17096105217933655, -0.028315693140029907, 0.010096178390085697, -0.04958370700478554, 0.003352926578372717, -0.06644797325134277, 0.019735202193260193, -0.03388429433107376, 0.008996130898594856, -0.02009434439241886, -0.08618813008069992, 0.08811068534851074, 0.10249514132738113, 0.07846277952194214, -0.27890726923942566, -0.06627105176448822, -0.01588202640414238, 0.019943663850426674, -0.004636147990822792, 0.04895554110407829, -0.05861157178878784, -0.07755891233682632, -0.01931183785200119, -0.032055068761110306, 0.1508222371339798, -0.031787652522325516, 0.03302391245961189, -0.1317082792520523, -0.018078558146953583, -0.005723212379962206, 0.050336338579654694, -0.19527585804462433, 0.24605390429496765, 0.0020159112755209208, 0.008772416040301323, -0.008561749011278152, -0.044392529875040054, 0.05571260303258896, 0.02237171120941639, 0.1663064807653427, 0.0006323985289782286, -0.04748477786779404, -0.08452274650335312, -0.051499754190444946, 0.09681772440671921, 0.06040066480636597, -0.06566203385591507, 0.06949575990438461, -0.04555588960647583, 0.02487722598016262, 0.039771176874637604, 0.1163230612874031, -0.08519209921360016, -0.05630447342991829, -0.020222550258040428, 0.012837645597755909, -0.029436426237225533, 0.027253583073616028, -0.04834328591823578, -0.00230967509560287, 0.188897043466568, -0.09278544783592224, -0.07283372431993484, -0.14754459261894226, 0.17685164511203766, 0.11510773748159409, -0.07062072306871414, 0.063358373939991, 0.0007518884376622736, -0.031568821519613266, 0.04718261584639549, -0.1804405152797699, 0.06344883888959885, -0.08023511618375778, 0.030185740441083908, -0.003505295841023326, 0.02280670404434204, 0.02976938523352146, 0.015803262591362, 0.03931993618607521, 0.051190514117479324, -0.08898928016424179, -0.08984269946813583, -0.009651989676058292, 0.01306087151169777, 0.0596698522567749, 0.13506647944450378, -0.11818736791610718, -0.028208184987306595, -0.010465244762599468, 0.06504771113395691, 0.15856990218162537, 0.08670129626989365, -0.06738109141588211, 0.0041035437025129795, 0.11169817298650742, -0.053423892706632614, -0.2939892113208771, -0.028595304116606712, -0.04864322766661644, 0.05315022170543671, -0.015318848192691803, -0.07573341578245163, 0.09873050451278687, -0.017544042319059372, -0.036863598972558975, -0.11198695003986359, -0.1124662384390831, -0.051046401262283325, 0.22682294249534607, 0.059804439544677734, 0.38951346278190613, -0.06173086166381836, -0.03674936294555664, -0.01750769466161728, -0.16181175410747528, 0.0640387088060379, 0.00645658141002059, 0.10409818589687347, -0.03935479745268822, -0.030810009688138962, 0.013407113961875439, -0.03945453092455864, 0.04382520169019699, 0.07891687005758286, 0.07999064028263092, -0.03576505184173584, -0.007150033954530954, 0.05379199981689453, 0.013476301915943623, 0.15008926391601562, 0.08086197823286057, 0.12464877963066101, -0.0788106620311737, -0.029534321278333664, -0.0671994611620903, 0.0725589171051979, 0.07013149559497833, -0.02998782880604267, -0.03922498971223831, 0.020142562687397003, 0.008864914998412132, 0.02445676177740097, 0.11139650642871857, -0.02465524524450302, -0.004493034444749355, 0.04366518557071686, 0.054502155631780624, -0.039883386343717575, -0.06431781500577927, 0.0017878103535622358, -0.030681759119033813, 0.09588600695133209, -0.12141899019479752, 0.06474357098340988, 0.13938604295253754, 0.0728878527879715, 0.07904131710529327, 0.12467367947101593, 0.005834502167999744, 0.009217801503837109, 0.07312231510877609, -0.1310526579618454, -0.002590365009382367, 0.0062270197086036205, -0.2033710777759552, -0.0006521020550280809, 0.11320420354604721, 0.18453918397426605, -0.08131662011146545, 0.005583268124610186, 0.012948202900588512, -0.0327775739133358, -0.029810581356287003, 0.1185709536075592, 0.12287573516368866, -0.005719451699405909, -0.1252547651529312, 0.13644875586032867, -0.0034536465536803007, -0.07566332817077637, 0.04198356345295906, -0.11587666720151901, -0.16774523258209229, -0.06923869252204895, 0.01339155063033104, 0.12430820614099503, -0.15036344528198242, -0.04307691752910614, -0.1881999373435974, -0.0761236697435379, 0.053112342953681946, 0.08786770701408386, 0.12447009980678558, 0.07120563089847565, -0.06818822026252747, -0.006435280665755272, -0.04155929014086723, 0.051003336906433105, 0.09024801850318909, -0.012863349169492722, -0.08335312455892563, 0.01523348968476057, -0.047259680926799774, 0.05536715313792229, -0.08411245793104172, -0.016799084842205048, -0.13856466114521027, 0.06720805168151855, -0.0538129098713398, -0.014318675734102726, -0.0534089170396328, -0.016674460843205452, 0.023584648966789246, -0.07857668399810791, -0.08185914903879166, 0.09595461189746857, -0.06487748771905899, -0.023455582559108734, 0.022970855236053467, 0.03081142157316208, -0.09784094989299774, -0.04289335757493973, 0.0875403881072998, -0.04837457463145256, 0.0664149820804596, 0.21856574714183807, -0.07773512601852417, 0.04579992592334747, -0.15400151908397675, -0.13487780094146729, 0.049296729266643524, 0.03059735894203186, 0.08083534240722656, -0.0361950509250164, 0.062339089810848236, 0.0662936419248581, 0.0013658751267939806, -0.0025529260747134686, 0.05638075992465019, -0.05439939349889755, -0.045087993144989014, -0.07668883353471756, -0.031945861876010895, -0.06183221936225891, -0.07258331030607224, 0.10985252261161804, 0.15345759689807892, 0.10611651092767715, -0.020938698202371597, 0.014004673808813095, -0.12560954689979553, 0.019085653126239777, 0.012732182629406452, -0.10069245100021362, -0.07343276590108871, -0.09222202003002167, 0.04387035593390465, -0.02074064128100872, 0.1502762734889984, -0.03625333309173584, 0.05831269547343254, 0.028562474995851517, 0.09113970398902893, 0.0385105274617672, 0.005584894213825464, 0.16699033975601196, 0.058238640427589417, -0.04061388224363327, -0.02071167342364788, 0.048372384160757065, 0.022986462339758873, -0.008936196565628052, 0.0014709579991176724, 0.1732836812734604, -0.007786605507135391, 0.07456684112548828, 0.04020638391375542, 0.023721473291516304, 0.027791257947683334, -0.041183579713106155, 0.0051489174365997314, 0.044825803488492966, 0.07747087627649307, -0.038778554648160934, 0.16749584674835205, 0.015210477635264397, 0.037609413266181946, -0.0071305944584310055, -0.03498123213648796, -0.10527265816926956, -0.04267815127968788, -0.11029116064310074, -0.12163206934928894, 0.02568093314766884, -0.09893898665904999, -0.08461298793554306, 0.1543172299861908, 0.036888718605041504, -0.012259021401405334, 0.07807651162147522, 0.03538717329502106, -0.09615297615528107, 0.05009603500366211, -0.04875628650188446, -0.053814008831977844, -0.003926421049982309, 0.029807869344949722, 0.07011454552412033, 0.07734300941228867, 0.012377338483929634, 0.0026194369420409203, 0.05022195726633072, 0.029680712148547173, -0.07299758493900299, -0.044509898871183395, -0.06926465034484863, 0.026817824691534042, -0.045565634965896606, 0.03451203182339668, 0.02782752737402916, -0.014192054979503155, 0.02840413711965084, 0.21898005902767181, -0.01620917208492756, -0.08312015235424042, -0.18404918909072876, 0.23008285462856293, 0.0852791965007782, 0.0649043619632721, -0.0046293060295283794, -0.045339833945035934, -0.0231457632035017, 0.26598262786865234, 0.1824505776166916, -0.08432630449533463, 0.002363852458074689, 0.03626875951886177, 0.0035324625205248594, 0.0443110391497612, -0.01557908020913601, 0.05048113688826561, 0.11962577700614929, -0.0251783300191164, -0.04763558506965637, -0.0741465613245964, -0.02774680219590664, -0.014615999534726143, -0.000885375018697232, 0.06894396990537643, -0.03547034040093422, -0.06452468037605286, 0.1138557717204094, -0.11690378934144974, -0.16278187930583954, -0.0042859334498643875, -0.10036996752023697, -0.08176207542419434, -0.061314601451158524, 0.006854790262877941, 0.060713641345500946, 0.051332514733076096, -0.08982378989458084, 0.05375780165195465, 0.010528404265642166, 0.03748730942606926, -0.11296583712100983, -0.06352920830249786, 0.052633076906204224, 0.005242037121206522, 0.007672118954360485, -0.02753640152513981, 0.039550431072711945, 0.08323967456817627, 0.019100207835435867, -0.10065089166164398, 0.0767166018486023, -0.02611156553030014, -0.02213672734797001, 0.051050398498773575, 0.006363440304994583, -0.009291338734328747, 0.009907455183565617, 0.03104078210890293, -0.10186946392059326, 0.023952338844537735, 0.07957427948713303, -0.018040841445326805, -0.11455974727869034, 0.03229518607258797, -0.06712117791175842, 0.10527107119560242, 0.11918973177671432, -0.04478808864951134, 0.029155952855944633, -0.06951295584440231, 0.0574994720518589, 0.0508681982755661, -0.082194022834301, -0.02130119875073433, -0.11225428432226181, -0.017157047986984253, -0.042967986315488815, -0.0505111962556839, -0.13665629923343658, -0.037980545312166214, -0.06770867854356766, -0.03325768932700157, 0.016537267714738846, 0.024904120713472366, 0.05507813021540642, 0.06875620782375336, -0.010089833289384842, -0.12768445909023285, 0.052991800010204315, 0.038793325424194336, -0.10909927636384964, -0.09670482575893402 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.5774 - Matthews Correlation: 0.5332 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.2347 | 1.0 | 535 | 0.5774 | 0.5332 | ### Framework versions - Transformers 4.11.0 - Pytorch 1.9.0+cu102 - Datasets 1.12.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["matthews_correlation"], "model-index": [{"name": "distilbert-base-uncased-finetuned-cola", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "glue", "type": "glue", "args": "cola"}, "metrics": [{"type": "matthews_correlation", "value": 0.533214904586951, "name": "Matthews Correlation"}]}]}]}
text-classification
BearThreat/distilbert-base-uncased-finetuned-cola
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-cola ====================================== This model is a fine-tuned version of distilbert-base-uncased on the glue dataset. It achieves the following results on the evaluation set: * Loss: 0.5774 * Matthews Correlation: 0.5332 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.11.0 * Pytorch 1.9.0+cu102 * Datasets 1.12.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.0\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ 67, 98, 4, 34 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.11.0\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3" ]
[ -0.10279300808906555, 0.10422883182764053, -0.0023008733987808228, 0.12261922657489777, 0.16534045338630676, 0.03359909728169441, 0.12659643590450287, 0.12829890847206116, -0.08497884124517441, 0.022294284775853157, 0.12140294164419174, 0.1582917422056198, 0.022134192287921906, 0.11519323289394379, -0.0502137765288353, -0.26404300332069397, -0.012186695821583271, 0.04754037782549858, -0.05425567179918289, 0.1342552900314331, 0.09264209866523743, -0.12076542526483536, 0.09056606888771057, 0.012744373641908169, -0.19401219487190247, -0.003862919518724084, -0.00020938878878951073, -0.053054917603731155, 0.147613063454628, 0.025319857522845268, 0.12458442896604538, 0.0005793149466626346, 0.08742702007293701, -0.19497738778591156, 0.010095656849443913, 0.046993423253297806, 0.004566281568259001, 0.09412289410829544, 0.045775920152664185, 0.0033470436464995146, 0.11568444222211838, -0.08128926157951355, 0.0550539493560791, 0.022559447214007378, -0.11553387343883514, -0.20870856940746307, -0.07989417016506195, 0.03725804015994072, 0.07807736098766327, 0.10570641607046127, -0.005402274429798126, 0.1170099675655365, -0.07754162698984146, 0.09205612540245056, 0.21837393939495087, -0.28596824407577515, -0.06569346785545349, 0.04403585568070412, 0.013159025460481644, 0.0443250872194767, -0.10069528967142105, -0.036820217967033386, 0.04614994302392006, 0.0527377650141716, 0.12735877931118011, -0.02957088127732277, -0.12057360261678696, 0.002893372904509306, -0.14039726555347443, -0.03417927771806717, 0.16918297111988068, 0.03965025767683983, -0.029187604784965515, -0.05353453382849693, -0.06129767373204231, -0.14558938145637512, -0.036445606499910355, -0.010892453603446484, 0.046457819640636444, -0.021294180303812027, -0.041284043341875076, -0.011890480294823647, -0.10891835391521454, -0.0626889243721962, -0.07784482836723328, 0.11049460619688034, 0.035646483302116394, 0.009138445369899273, -0.026921967044472694, 0.11201941967010498, -0.0046372124925255775, -0.1238493025302887, 0.02372582070529461, 0.02326519414782524, 0.01263529621064663, -0.039672594517469406, -0.05352193862199783, -0.0614241287112236, 0.011294220574200153, 0.12980283796787262, -0.0487070307135582, 0.04137416183948517, 0.04877876117825508, 0.0497160330414772, -0.09099246561527252, 0.19253046810626984, -0.03437615931034088, -0.028195422142744064, 0.010281330905854702, 0.04663441330194473, 0.01978146843612194, -0.012178588658571243, -0.1256365031003952, 0.005771835800260305, 0.0895320400595665, 0.00809482205659151, -0.0604364350438118, 0.0735849067568779, -0.055712416768074036, -0.02572866529226303, 0.0045046028681099415, -0.09240616858005524, 0.022508323192596436, -0.0006571399280801415, -0.07055362313985825, -0.02105095610022545, 0.03662649542093277, 0.016043853014707565, -0.02029290609061718, 0.10822336375713348, -0.0873841643333435, 0.02846844308078289, -0.09276894479990005, -0.10831838101148605, 0.01874585822224617, -0.10543303936719894, 0.021020647138357162, -0.09493139386177063, -0.18726618587970734, -0.01671597920358181, 0.06202967092394829, -0.024370985105633736, -0.06186693161725998, -0.054573509842157364, -0.06852193921804428, 0.0132385129109025, -0.009420176967978477, 0.11737555265426636, -0.06413032859563828, 0.09143359214067459, 0.018750140443444252, 0.06006000190973282, -0.04435410350561142, 0.059814371168613434, -0.10299103707075119, 0.016587624326348305, -0.1521376520395279, 0.041767753660678864, -0.050863973796367645, 0.06855813413858414, -0.08281469345092773, -0.10343360155820847, 0.008961272425949574, -0.004794985521584749, 0.061744049191474915, 0.09288926422595978, -0.1886129379272461, -0.07540222257375717, 0.15604326128959656, -0.07296520471572876, -0.12155035883188248, 0.12066496908664703, -0.06019167602062225, 0.05720491707324982, 0.05716344341635704, 0.1779988706111908, 0.0813363641500473, -0.07673565298318863, 0.0016506616957485676, 0.025558333843946457, 0.050469979643821716, -0.0673050507903099, 0.06937256455421448, 0.004316362552344799, 0.019244709983468056, 0.03528444841504097, -0.028316481038928032, 0.06305693835020065, -0.08607903122901917, -0.09872300922870636, -0.04146211966872215, -0.0815219134092331, 0.041617073118686676, 0.07491133362054825, 0.0683816596865654, -0.09047376364469528, -0.07821202278137207, 0.05091281607747078, 0.0828106701374054, -0.05764056742191315, 0.024508841335773468, -0.04992995783686638, 0.07495249062776566, -0.02769765816628933, -0.02306230552494526, -0.18205304443836212, -0.03828331083059311, 0.007995646446943283, 0.00005913667700951919, 0.01720590889453888, 0.02866414189338684, 0.06020102649927139, 0.06021204590797424, -0.048094894737005234, -0.01680595614016056, -0.031857576221227646, 0.0013455442385748029, -0.12780693173408508, -0.19138456881046295, -0.030924102291464806, -0.024158144369721413, 0.15946686267852783, -0.2055101841688156, 0.048688121140003204, -0.015790415927767754, 0.07046819478273392, 0.012541000731289387, -0.006207108031958342, -0.03763170912861824, 0.07250089198350906, -0.045843709260225296, -0.054314713925123215, 0.08063612878322601, 0.018923748284578323, -0.08989064395427704, -0.04833566024899483, -0.0965856984257698, 0.1516607701778412, 0.12742529809474945, -0.10575906932353973, -0.07663419842720032, -0.020456181839108467, -0.06784219294786453, -0.03364754095673561, -0.049054257571697235, 0.024748073890805244, 0.18800413608551025, -0.003906234400346875, 0.1503865122795105, -0.06733424961566925, -0.04328780248761177, 0.01743065007030964, -0.03750794753432274, 0.01721475087106228, 0.12696918845176697, 0.1366567462682724, -0.06061924248933792, 0.15456748008728027, 0.14683355391025543, -0.08989774435758591, 0.14343981444835663, -0.041477955877780914, -0.06381718069314957, -0.01670766808092594, -0.03213339298963547, -0.010730238631367683, 0.10110044479370117, -0.15134243667125702, 0.0024292629677802324, 0.03463253751397133, 0.01776096597313881, 0.025880740955471992, -0.22553198039531708, -0.03901065140962601, 0.0343007892370224, -0.04298141226172447, -0.00342431734316051, -0.008657626807689667, 0.007458264008164406, 0.10115987062454224, 0.0014916773652657866, -0.08629979193210602, 0.04003053158521652, 0.0021486219484359026, -0.0840182974934578, 0.21549959480762482, -0.08348015695810318, -0.17473360896110535, -0.13285177946090698, -0.07171723991632462, -0.048058900982141495, 0.0013497670879587531, 0.06628609448671341, -0.08669362962245941, -0.031177349388599396, -0.07344099134206772, 0.02311423234641552, 0.009616542607545853, 0.02361954003572464, 0.004558672662824392, 0.0050300173461437225, 0.06426215916872025, -0.11129877716302872, -0.016002114862203598, -0.05655830353498459, -0.04465613514184952, 0.04503943398594856, 0.031963977962732315, 0.11155882477760315, 0.15488503873348236, -0.014235367067158222, 0.011503051966428757, -0.02877158485352993, 0.23828376829624176, -0.0601540245115757, -0.015435434877872467, 0.14454734325408936, -0.01253626961261034, 0.05278300866484642, 0.119680255651474, 0.07310093939304352, -0.07729794830083847, 0.004295294638723135, 0.035034868866205215, -0.03734712675213814, -0.22829604148864746, -0.059667374938726425, -0.05686764046549797, 0.009799269959330559, 0.09267961233854294, 0.0251909252256155, 0.029063742607831955, 0.07218023389577866, 0.041536036878824234, 0.07774540036916733, -0.03989090397953987, 0.055256448686122894, 0.13252684473991394, 0.03289283812046051, 0.12513934075832367, -0.04500951990485191, -0.06360883265733719, 0.04493417963385582, -0.009458308108150959, 0.22453121840953827, 0.005118488799780607, 0.12870241701602936, 0.061843764036893845, 0.16245518624782562, -0.005487077869474888, 0.07769262790679932, -0.009705791249871254, -0.0342765748500824, -0.018496612086892128, -0.03875783458352089, -0.03990678861737251, 0.026431413367390633, -0.06703755259513855, 0.06122253090143204, -0.11839199811220169, 0.015610144473612309, 0.05896832048892975, 0.24873103201389313, 0.03569231554865837, -0.3219321370124817, -0.0998094230890274, 0.0035814165603369474, -0.03238970413804054, -0.022894619032740593, 0.026992443948984146, 0.09527654945850372, -0.1011265441775322, 0.02813813090324402, -0.07541372627019882, 0.0972311720252037, -0.05420665442943573, 0.04744546487927437, 0.08413779735565186, 0.09042283892631531, 0.012764123268425465, 0.09374187886714935, -0.28325745463371277, 0.2730015218257904, -0.00032767787342891097, 0.05594548583030701, -0.0787319615483284, 0.010795503854751587, 0.04270555078983307, 0.06342089921236038, 0.08118455111980438, -0.01109253615140915, -0.026944704353809357, -0.18373681604862213, -0.07208176702260971, 0.028036510571837425, 0.061568934470415115, -0.038567040115594864, 0.0838024914264679, -0.032913703471422195, 0.007837118580937386, 0.07168308645486832, 0.001422868575900793, -0.049012333154678345, -0.10880114883184433, -0.005403253715485334, 0.02389039844274521, -0.06047046184539795, -0.06088842824101448, -0.11968328058719635, -0.12497910112142563, 0.15882718563079834, -0.030894389376044273, -0.04107281193137169, -0.10898321866989136, 0.08531053364276886, 0.06234687194228172, -0.0898977667093277, 0.04522910341620445, -0.00009666448750067502, 0.080811507999897, 0.021980052813887596, -0.07439424097537994, 0.099859319627285, -0.07583875209093094, -0.1575811803340912, -0.06650251895189285, 0.10620861500501633, 0.03146900236606598, 0.06534453481435776, -0.011564468964934349, 0.008772975765168667, -0.048598118126392365, -0.0902535691857338, 0.016531016677618027, 0.00885729305446148, 0.07965926080942154, 0.01828465424478054, -0.07710863649845123, 0.009418398141860962, -0.05949688330292702, -0.03278909996151924, 0.20967362821102142, 0.21516357362270355, -0.10291747748851776, 0.0258738175034523, 0.022573521360754967, -0.07323221117258072, -0.2022797018289566, 0.030979884788393974, 0.05708850175142288, 0.008206317201256752, 0.04069980978965759, -0.1808328777551651, 0.13968941569328308, 0.10828717797994614, -0.014858600683510303, 0.10581323504447937, -0.32038867473602295, -0.12251647561788559, 0.1355663537979126, 0.13295039534568787, 0.1013026162981987, -0.13052129745483398, -0.02210475131869316, -0.01950845867395401, -0.1369461566209793, 0.11956405639648438, -0.0884537324309349, 0.12002845108509064, -0.03388483077287674, 0.08285260945558548, 0.001898106187582016, -0.05794650688767433, 0.11994529515504837, 0.030726028606295586, 0.09133085608482361, -0.05965253338217735, -0.034702520817518234, 0.03181315213441849, -0.044482771307229996, 0.035291217267513275, -0.09161004424095154, 0.03116159699857235, -0.10661076009273529, -0.025455055758357048, -0.06529095023870468, 0.04694397747516632, -0.04244722053408623, -0.06938926875591278, -0.03732076287269592, 0.026195749640464783, 0.049562789499759674, -0.009380510076880455, 0.1228184774518013, 0.027801858261227608, 0.14086660742759705, 0.10167163610458374, 0.06906361877918243, -0.0703466385602951, -0.07843361794948578, -0.027152519673109055, -0.011653848923742771, 0.04933365434408188, -0.13486483693122864, 0.02282138727605343, 0.15230783820152283, 0.019655762240290642, 0.15173035860061646, 0.08230976015329361, -0.017933087423443794, 0.0009208131232298911, 0.056495632976293564, -0.16588246822357178, -0.08880606293678284, -0.013872046954929829, -0.06536604464054108, -0.12021529674530029, 0.041427262127399445, 0.09501462429761887, -0.06666935235261917, -0.006955535616725683, -0.004348443821072578, 0.0151974530890584, -0.04756782203912735, 0.1831301748752594, 0.0623435378074646, 0.04645615071058273, -0.09822893142700195, 0.07191681116819382, 0.04912403225898743, -0.07163060456514359, 0.003510675858706236, 0.0740644559264183, -0.08810889720916748, -0.05513014271855354, 0.06612610816955566, 0.19047507643699646, -0.049941934645175934, -0.04589489847421646, -0.1407354176044464, -0.1230883002281189, 0.0788184404373169, 0.1384146362543106, 0.12017875909805298, 0.012202669866383076, -0.06910410523414612, 0.00011647997598629445, -0.10836810618638992, 0.10493104159832001, 0.05061253905296326, 0.06310340017080307, -0.1427597999572754, 0.14172016084194183, 0.017554597929120064, 0.05106544867157936, -0.020379748195409775, 0.02426968701183796, -0.09782145172357559, 0.0054598236456513405, -0.09977550804615021, -0.011416332796216011, -0.03453630581498146, 0.011789224110543728, -0.005290700122714043, -0.04711631312966347, -0.05515369400382042, 0.00996152963489294, -0.10625007748603821, -0.02359815686941147, 0.024840328842401505, 0.068723164498806, -0.10809236019849777, -0.03767780214548111, 0.028086870908737183, -0.0619608610868454, 0.07794138789176941, 0.04473499208688736, 0.01604904606938362, 0.04824193939566612, -0.13489893078804016, 0.017268406227231026, 0.0738854929804802, 0.03188799321651459, 0.0642159953713417, -0.09786971658468246, -0.007522415369749069, -0.0058600325137376785, 0.03803698346018791, 0.019768357276916504, 0.07885270565748215, -0.141421839594841, 0.00413302518427372, -0.023479042574763298, -0.08234965801239014, -0.06777331978082657, 0.02591903880238533, 0.09028016030788422, 0.021436871960759163, 0.20136193931102753, -0.07637091726064682, 0.05275068059563637, -0.21532082557678223, 0.005906919948756695, -0.008555068634450436, -0.10748061537742615, -0.10404549539089203, -0.07182567566633224, 0.054183583706617355, -0.057622455060482025, 0.1523776799440384, 0.04857052490115166, 0.02132020890712738, 0.024673225358128548, -0.007100203074514866, 0.014666268602013588, 0.011156144551932812, 0.18934164941310883, 0.030911331996321678, -0.03486789017915726, 0.057980574667453766, 0.04205835983157158, 0.10455138236284256, 0.11109081655740738, 0.2013447880744934, 0.14122328162193298, -0.005025472957640886, 0.09158527851104736, 0.041404105722904205, -0.05894056707620621, -0.1603473722934723, 0.04840333014726639, -0.0365777313709259, 0.10976921021938324, -0.019636502489447594, 0.21808229386806488, 0.05810067430138588, -0.17061612010002136, 0.0480346754193306, -0.05174854397773743, -0.08691450208425522, -0.11344516277313232, -0.05299512296915054, -0.07929187268018723, -0.12746861577033997, -0.005657556466758251, -0.11669203639030457, -0.002440051408484578, 0.12731194496154785, 0.0033397264778614044, -0.02767563797533512, 0.1544724553823471, 0.004971806425601244, 0.02212688699364662, 0.0561097227036953, 0.01243639551103115, -0.03465341404080391, -0.13249556720256805, -0.05960051715373993, -0.017217306420207024, -0.00554075138643384, 0.03260444477200508, -0.06053624674677849, -0.03692072629928589, 0.031608641147613525, -0.023486804217100143, -0.0933631956577301, 0.004544257186353207, 0.012821498326957226, 0.05403223633766174, 0.046100761741399765, 0.010908681899309158, 0.020007479935884476, -0.0029331515543162823, 0.20085826516151428, -0.07132938504219055, -0.06852628290653229, -0.10697580873966217, 0.23277424275875092, 0.03194531798362732, -0.02219802513718605, 0.03543568402528763, -0.06592653691768646, 0.0026551843620836735, 0.24856656789779663, 0.21759366989135742, -0.08385965973138809, -0.007574681658297777, 0.016771350055933, -0.009251066483557224, -0.021876318380236626, 0.10048367828130722, 0.14276152849197388, 0.05581008642911911, -0.09224972873926163, -0.04860039800405502, -0.05972401052713394, -0.017476698383688927, -0.03913731873035431, 0.07115549594163895, 0.04579627886414528, 0.007644144352525473, -0.0357162281870842, 0.054385777562856674, -0.07021918147802353, -0.09403801709413528, 0.055542148649692535, -0.21703816950321198, -0.16880258917808533, -0.012428735382854939, 0.09944067895412445, 0.0026170213241130114, 0.06020832434296608, -0.03112715296447277, -0.0035619225818663836, 0.09516675025224686, -0.02084827609360218, -0.09732617437839508, -0.06594754010438919, 0.08769720792770386, -0.10996295511722565, 0.2220596820116043, -0.04684586822986603, 0.055138155817985535, 0.12469585984945297, 0.06985293328762054, -0.07057791203260422, 0.06420459598302841, 0.04308589920401573, -0.03968115523457527, 0.02667825296521187, 0.07038924843072891, -0.03575694561004639, 0.06093055382370949, 0.04917791485786438, -0.1410796195268631, 0.018446387723088264, -0.04733744263648987, -0.06864605844020844, -0.04483827203512192, -0.026410982012748718, -0.062117066234350204, 0.13238827884197235, 0.21513433754444122, -0.026666980236768723, -0.01049641240388155, -0.07016243040561676, 0.010036715306341648, 0.05356132611632347, 0.022390590980648994, -0.05623801052570343, -0.21058981120586395, 0.01666071079671383, 0.03911347687244415, -0.019149182364344597, -0.24533239006996155, -0.1019408255815506, 0.0014413193566724658, -0.07278291881084442, -0.09514392167329788, 0.0743955448269844, 0.08408370614051819, 0.050093408674001694, -0.0573832243680954, -0.039159730076789856, -0.07618719339370728, 0.14587362110614777, -0.14390549063682556, -0.09243766963481903 ]
null
null
transformers
# Cartman Southpark DialoGPT2 small 18 epochs
{"tags": ["conversational"]}
text-generation
Bee-Garbs/DialoGPT-real-cartman-small
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2022-03-02T23:29:04+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Cartman Southpark DialoGPT2 small 18 epochs
[ "# Cartman Southpark DialoGPT2 small 18 epochs" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Cartman Southpark DialoGPT2 small 18 epochs" ]
[ 51, 16 ]
[ "passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Cartman Southpark DialoGPT2 small 18 epochs" ]
[ -0.03801896050572395, 0.05976199358701706, -0.004420805256813765, 0.0354803130030632, 0.13204607367515564, -0.010659758001565933, 0.15473578870296478, 0.12925772368907928, 0.021664133295416832, 0.006697777193039656, 0.15537017583847046, 0.15116840600967407, 0.005074323154985905, 0.05653480812907219, -0.10336565971374512, -0.2762875556945801, 0.016603922471404076, 0.04392891377210617, 0.030449585989117622, 0.12571220099925995, 0.042985573410987854, -0.07893843948841095, 0.06916652619838715, -0.02962460368871689, -0.12722043693065643, -0.00235893321223557, 0.04747126251459122, -0.12353271245956421, 0.11360497772693634, 0.017637524753808975, -0.009888185188174248, 0.024148166179656982, -0.012303204275667667, -0.04663919284939766, 0.033417344093322754, -0.014548859558999538, 0.005408284720033407, 0.05350074917078018, 0.006683171261101961, -0.03801283240318298, 0.10894924402236938, 0.10994327068328857, 0.016182249411940575, 0.030872080475091934, -0.14163780212402344, 0.001143673318438232, -0.06578238308429718, 0.033933378756046295, 0.10194183140993118, 0.1004478707909584, -0.006477124057710171, 0.1634853184223175, -0.06082307919859886, 0.08950433135032654, 0.1312473565340042, -0.35474735498428345, -0.04297946020960808, 0.12628906965255737, 0.057985588908195496, 0.09499642997980118, -0.09823393821716309, 0.08189400285482407, 0.011650756932795048, 0.014217516407370567, 0.0065827639773488045, -0.08848654478788376, -0.1233687624335289, 0.04605322703719139, -0.1138770803809166, -0.06200525164604187, 0.23932777345180511, -0.04018653556704521, 0.0662153959274292, -0.05646659433841705, -0.11963457614183426, -0.0744612067937851, -0.0723675936460495, -0.012701457366347313, -0.06718717515468597, 0.040419504046440125, -0.07126475125551224, -0.10377755016088486, -0.11298944056034088, -0.029690822586417198, -0.1705193817615509, 0.2320815622806549, 0.016227830201387405, 0.04871544986963272, -0.1869216114282608, 0.044623490422964096, 0.023304123431444168, -0.06880301237106323, 0.021134715527296066, -0.04380868002772331, 0.018557574599981308, 0.002860989887267351, -0.031699199229478836, -0.015344868414103985, 0.06650512665510178, 0.1903972029685974, 0.0015498852590098977, 0.01772862859070301, -0.029410546645522118, 0.044031593948602676, 0.022679448127746582, 0.10851704329252243, 0.0313730463385582, -0.0849861353635788, 0.0695938989520073, -0.055900827050209045, 0.05683009698987007, -0.06296421587467194, -0.1680404096841812, -0.023074399679899216, 0.04467463865876198, 0.06198614463210106, 0.027441274374723434, 0.10343295335769653, 0.05666138976812363, -0.033864445984363556, 0.049790650606155396, -0.037674084305763245, -0.03950730711221695, -0.0015576519072055817, -0.007740057073533535, 0.08877301216125488, -0.04501187056303024, 0.023835839703679085, -0.10298047214746475, 0.006938400212675333, -0.0768207311630249, -0.008054407313466072, 0.03192680701613426, -0.04350782558321953, 0.008214449509978294, -0.029867080971598625, -0.014199896715581417, -0.1606203317642212, -0.20307861268520355, 0.009221001528203487, -0.05165981873869896, -0.03299983963370323, -0.11793233454227448, -0.056869130581617355, -0.02166558988392353, 0.06380141526460648, -0.060845013707876205, -0.026055991649627686, -0.05066637322306633, 0.09042712301015854, -0.03160744532942772, 0.09175948798656464, -0.06648806482553482, 0.06666587293148041, -0.10979195684194565, -0.04328392073512077, -0.098518967628479, 0.0553763173520565, 0.004236217122524977, 0.04392003268003464, 0.009929278865456581, 0.017223868519067764, -0.12296482920646667, 0.051481328904628754, -0.05446361377835274, 0.21302448213100433, -0.07000406831502914, -0.10998866707086563, 0.22589965164661407, -0.03310592100024223, -0.12031660228967667, 0.1411912441253662, 0.0251814853399992, -0.006531921215355396, 0.11160656809806824, 0.2388318032026291, -0.022744562476873398, -0.005604930687695742, 0.060160789638757706, 0.0787024050951004, -0.06667610257863998, -0.07986974716186523, 0.04234995320439339, -0.031642284244298935, -0.016600092872977257, 0.0169367752969265, 0.060562558472156525, 0.0724620446562767, -0.04899570345878601, 0.021793784573674202, 0.004538987763226032, -0.02816912904381752, 0.14097975194454193, 0.01575872115790844, 0.14214074611663818, -0.0747281163930893, -0.05141931027173996, -0.011954362504184246, 0.008966387249529362, -0.01566525176167488, 0.034155286848545074, -0.06521972268819809, 0.1034608781337738, -0.015046876855194569, 0.038059771060943604, -0.14954057335853577, -0.09462365508079529, -0.011044076643884182, 0.21130827069282532, 0.06375677138566971, 0.11009039729833603, 0.06470327824354172, -0.04223281517624855, -0.014071411453187466, 0.04464279115200043, 0.1282864511013031, -0.01994531974196434, -0.13396945595741272, -0.049286745488643646, 0.06693888455629349, -0.04370695725083351, 0.07299485057592392, -0.02455834113061428, 0.026464838534593582, 0.025577440857887268, 0.07343374192714691, -0.027507346123456955, -0.00932000670582056, 0.024587389081716537, -0.025093968957662582, -0.08438707888126373, -0.010348155163228512, 0.09166933596134186, 0.021145856007933617, -0.02181960828602314, 0.25235071778297424, -0.24378322064876556, 0.06035597249865532, 0.18937624990940094, -0.1827189326286316, 0.007977129891514778, -0.07561288774013519, -0.021717600524425507, -0.0016814405098557472, 0.009458750486373901, -0.07573439180850983, 0.20303942263126373, -0.008720500394701958, 0.20124541223049164, -0.03513109311461449, -0.06880620867013931, -0.04251524433493614, -0.021522173658013344, -0.0015762788243591785, 0.07971281558275223, 0.13102319836616516, -0.12701648473739624, 0.1761181801557541, 0.14731067419052124, 0.07058539986610413, 0.20134414732456207, 0.020864900201559067, -0.010639489628374577, 0.07169090211391449, -0.037561364471912384, -0.060155585408210754, -0.048849500715732574, -0.236373633146286, -0.049567461013793945, 0.07255736738443375, 0.04021006077528, 0.08749359101057053, -0.09182467311620712, -0.013080397620797157, -0.015383831225335598, -0.002850604709237814, 0.09553755074739456, 0.11029371619224548, 0.05536029860377312, 0.14390695095062256, 0.01516753900796175, -0.06336310505867004, 0.061966393142938614, 0.0316624790430069, -0.08080220967531204, 0.16346058249473572, -0.10359489917755127, -0.30614879727363586, -0.10708119720220566, -0.15656369924545288, -0.048133183270692825, 0.0459187813103199, 0.09655475616455078, -0.09228344261646271, 0.013293917290866375, -0.011326534673571587, 0.1008339524269104, -0.08266513794660568, 0.02306211367249489, 0.04416162893176079, -0.017229683697223663, -0.15335209667682648, -0.07193151861429214, -0.06409896910190582, -0.07426658272743225, -0.07550451904535294, 0.11205165833234787, -0.08120759576559067, -0.006951119750738144, 0.21886873245239258, 0.02641412988305092, 0.05380775406956673, -0.0696910098195076, 0.16938576102256775, -0.0721060112118721, 0.016333427280187607, 0.24674378335475922, -0.0048055327497422695, 0.04121362417936325, 0.06249556690454483, -0.022622311487793922, -0.07730455696582794, 0.016591835767030716, -0.0036808412987738848, -0.11211453378200531, -0.2078913152217865, -0.10250400751829147, -0.13143126666545868, 0.16808432340621948, 0.05307420715689659, 0.03401150181889534, 0.07725632935762405, 0.04219719022512436, -0.08438455313444138, 0.06033894792199135, 0.016333812847733498, 0.06735420972108841, 0.30327266454696655, -0.03043103590607643, 0.13269038498401642, -0.004174739588052034, -0.09643331170082092, 0.10526596754789352, -0.0044736252166330814, 0.04958690330386162, 0.08125200867652893, 0.08518556505441666, -0.038821592926979065, 0.006301112473011017, 0.11875828355550766, 0.019236695021390915, 0.0813836082816124, -0.04456700384616852, -0.017478445544838905, -0.02267499640583992, -0.07990198582410812, 0.03251529112458229, 0.07093548774719238, -0.14288300275802612, -0.03797364607453346, 0.03685828670859337, 0.03337162360548973, 0.1092650443315506, 0.0981798768043518, -0.17111369967460632, -0.06360869854688644, 0.054814599454402924, -0.09339143335819244, -0.08247419446706772, 0.14472989737987518, 0.0326685905456543, -0.1354847252368927, 0.01999303698539734, -0.03389282897114754, 0.09927064925432205, -0.08860981464385986, 0.08102106302976608, -0.10997620970010757, -0.09325249493122101, 0.002540505025535822, 0.04730476811528206, -0.16513864696025848, 0.1974286437034607, -0.017021821811795235, -0.07424423843622208, -0.08613546937704086, -0.011230656877160072, 0.04113485664129257, 0.031245563179254532, 0.057824429124593735, 0.014639248140156269, -0.06892375648021698, 0.000025527997422614135, -0.06600498408079147, -0.0063890558667480946, 0.10244958847761154, -0.08481299132108688, 0.017718149349093437, -0.06752520054578781, -0.0033448170870542526, -0.03598468378186226, -0.1146702691912651, 0.029983624815940857, -0.11347044259309769, 0.07760434597730637, 0.08407667279243469, 0.003197189886122942, 0.028409568592905998, -0.03529622033238411, -0.1470957249403, 0.2084556221961975, 0.025565490126609802, -0.0811678022146225, -0.10642722249031067, 0.020156631246209145, 0.04714252054691315, -0.0692865252494812, 0.04398972913622856, -0.04597044363617897, 0.04852161929011345, -0.05542163923382759, -0.18191541731357574, 0.11378180980682373, -0.09829974174499512, -0.0916285365819931, -0.010769441723823547, 0.23291200399398804, -0.0702303797006607, 0.035365957766771317, 0.018530145287513733, 0.04231758043169975, -0.06810146570205688, -0.06994068622589111, 0.061281438916921616, -0.05812758579850197, 0.014034505002200603, 0.011035284027457237, 0.030237460508942604, -0.04854908213019371, -0.017781538888812065, -0.040132515132427216, 0.31670960783958435, 0.19392980635166168, -0.04927733168005943, 0.15052105486392975, 0.10642065107822418, -0.019981257617473602, -0.30612921714782715, -0.10725262761116028, -0.1042531281709671, -0.048453669995069504, 0.010242893360555172, -0.11821848899126053, 0.0709344819188118, 0.032719869166612625, -0.013841727748513222, 0.07491588592529297, -0.344280868768692, -0.10030783712863922, 0.15798550844192505, -0.026922903954982758, 0.40882495045661926, -0.12142082303762436, -0.08008706569671631, -0.002979161450639367, -0.15089541673660278, 0.11066247522830963, 0.012562589719891548, 0.1087278202176094, -0.049622200429439545, 0.1796775907278061, 0.04664990305900574, 0.0003214164753444493, 0.04576259106397629, -0.018582690507173538, -0.048181235790252686, -0.09081070870161057, -0.10655500739812851, 0.06276312470436096, 0.0023219254799187183, -0.018694955855607986, -0.02804109826683998, 0.017311377450823784, -0.12115788459777832, -0.01471257209777832, -0.07029148191213608, 0.03335835039615631, 0.04023413360118866, -0.0747925341129303, -0.014721452258527279, -0.07099001109600067, -0.05713127180933952, -0.005729072727262974, 0.21618172526359558, -0.06495992094278336, 0.1583952158689499, 0.09267831593751907, 0.10571005195379257, -0.10837946087121964, -0.02050008811056614, -0.0954030305147171, -0.03621964156627655, 0.022443529218435287, -0.10315310955047607, 0.01817995123565197, 0.08595524728298187, 0.03345276042819023, 0.02473362348973751, 0.0665520653128624, 0.025423625484108925, 0.03506452962756157, 0.040079060941934586, -0.24995960295200348, -0.07404724508523941, -0.06406914442777634, 0.0052247243002057076, 0.0853566974401474, 0.09074296802282333, 0.2057773768901825, 0.00016591687744949013, -0.05499379336833954, -0.010153146460652351, 0.009819120168685913, 0.044487711042165756, 0.06337011605501175, -0.037934064865112305, 0.04345173016190529, -0.13539911806583405, 0.07553473860025406, 0.02443145401775837, -0.0973886176943779, 0.024996403604745865, 0.23037917912006378, -0.11275333166122437, -0.11431992053985596, -0.08065570145845413, 0.0108861792832613, -0.11228964477777481, 0.047091465443372726, -0.06846176832914352, -0.10931649059057236, 0.08176464587450027, 0.1582752764225006, 0.04682060703635216, 0.06788816303014755, -0.06598465144634247, 0.002510702470317483, -0.022345848381519318, 0.017114154994487762, -0.016865644603967667, 0.00707204220816493, -0.07464158535003662, 0.08574378490447998, -0.04311821609735489, 0.11250080913305283, -0.08779222518205643, -0.09041276574134827, -0.15011277794837952, 0.023758655413985252, -0.1169465184211731, -0.07431154698133469, -0.11727376282215118, -0.06606587022542953, -0.030507665127515793, -0.017428359016776085, -0.0903143659234047, -0.03617720305919647, -0.11614803224802017, 0.01528206653892994, -0.04114307463169098, 0.027488337829709053, -0.04885794594883919, 0.038950420916080475, 0.06727353483438492, -0.028178617358207703, 0.17337079346179962, 0.15929405391216278, -0.045310795307159424, 0.08608875423669815, -0.08720697462558746, -0.07858388870954514, 0.05862298980355263, 0.03086838871240616, 0.06107911095023155, 0.12469968944787979, 0.0017599432030692697, 0.04148050770163536, 0.07167825102806091, 0.08360740542411804, 0.04283226653933525, -0.10079680383205414, 0.025417523458600044, -0.041391726583242416, -0.10673359781503677, -0.05160275474190712, -0.03496137633919716, 0.04806656762957573, 0.023092739284038544, 0.0706339180469513, -0.0638715922832489, 0.07609330117702484, -0.08936592191457748, 0.03133464604616165, -0.004875361919403076, -0.18096542358398438, -0.05774746462702751, -0.05049695819616318, 0.04688596352934837, 0.032662276178598404, 0.1581418663263321, 0.033876679837703705, -0.018444783985614777, 0.06895804405212402, 0.1151648610830307, 0.06926490366458893, -0.016052471473813057, 0.13530154526233673, 0.10497846454381943, -0.08797188848257065, -0.04111258313059807, 0.06257515400648117, 0.0728757232427597, -0.011954368092119694, 0.2051677703857422, -0.019851969555020332, -0.013594087213277817, 0.07441212981939316, -0.02583632618188858, -0.0008610764052718878, -0.17966850101947784, -0.11864454299211502, -0.039393100887537, 0.06688986718654633, -0.07207600772380829, 0.14841414988040924, 0.19084285199642181, -0.018788665533065796, -0.0033852302003651857, -0.06363137811422348, -0.026942407712340355, -0.16322225332260132, -0.12672387063503265, -0.07213901728391647, -0.13968051970005035, 0.013037120923399925, -0.0970701277256012, 0.05652080476284027, 0.10461284965276718, 0.07967238873243332, -0.041564345359802246, 0.05568290129303932, 0.01050429418683052, -0.11480338126420975, 0.05694813281297684, -0.03672285005450249, 0.057869333773851395, -0.0575094036757946, -0.05611572414636612, -0.07329002767801285, -0.01951269619166851, 0.012338193133473396, 0.005135842598974705, -0.03481090068817139, 0.00045728430268354714, -0.15353508293628693, -0.08947348594665527, -0.03841143101453781, 0.013906252570450306, 0.021787051111459732, 0.08589958399534225, 0.030781716108322144, -0.009030754677951336, 0.013709391467273235, 0.28624415397644043, -0.08927236497402191, -0.07002830505371094, -0.01776399277150631, 0.11311198025941849, 0.016294052824378014, 0.08120489865541458, -0.020852575078606606, -0.008082273416221142, -0.0524626262485981, 0.3310869634151459, 0.22827288508415222, -0.05833674222230911, 0.030938901007175446, 0.014251232147216797, 0.02356916107237339, 0.06772230565547943, 0.1303299516439438, 0.09042613953351974, 0.21415556967258453, -0.047487422823905945, -0.12414095550775528, -0.00876687839627266, -0.029657678678631783, -0.13304691016674042, 0.08653160184621811, 0.01811901293694973, -0.04474276676774025, -0.0025681022088974714, 0.07835495471954346, -0.188863143324852, 0.06501562148332596, -0.16901375353336334, -0.21706947684288025, -0.07654283195734024, 0.015638019889593124, 0.11284962296485901, 0.028363436460494995, 0.0890401303768158, 0.015179600566625595, -0.05911305174231529, -0.05003209412097931, 0.011310325004160404, -0.22484485805034637, -0.002289463998749852, 0.08060736209154129, -0.08991603553295135, 0.05351730436086655, -0.07157213985919952, 0.06501778215169907, 0.08579091727733612, 0.06968111544847488, -0.03134533390402794, 0.036854151636362076, 0.03846704959869385, -0.06816376000642776, 0.03590710833668709, 0.09387128800153732, -0.004896252881735563, 0.029100915417075157, 0.07859162986278534, -0.09735769778490067, 0.026428300887346268, 0.012826011516153812, -0.012033192440867424, 0.006540622096508741, 0.0005814231699332595, -0.038366641849279404, 0.10767140239477158, 0.04137234017252922, -0.0353623665869236, -0.02304653264582157, -0.020124465227127075, -0.06871192902326584, -0.026181038469076157, -0.02111021988093853, -0.10455244034528732, -0.19932517409324646, -0.10345723479986191, 0.055958766490221024, -0.006884515751153231, -0.20012696087360382, -0.001251396955922246, -0.14060692489147186, 0.06354861706495285, -0.14973875880241394, 0.10353054851293564, 0.08632439374923706, -0.03733526170253754, 0.0018371616024523973, 0.03748602792620659, 0.024045569822192192, 0.11074743419885635, -0.165806382894516, -0.07138963788747787 ]