rlaorrn commited on
Commit
966d24c
1 Parent(s): 9760614

Training in progress, step 500

Browse files
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7623790a033170c16e7a6620cf2811e4b47cf15059338f045e9eb1d02c9692c3
3
  size 377611120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:961dc10472077d0896da30b8695710d2c78ff5b4aa1cdc2ae8b68012760dc1a7
3
  size 377611120
runs/May26_14-12-33_142d81d78b72/events.out.tfevents.1716732784.142d81d78b72.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8aeb8da49cecff2e3dcb6ae165369cebc8512e825be136aceb15642afc851062
3
+ size 6848
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16c44ddbf94d57ba0bb11a0f097172952efe63cc2b9e9b68519186850c66a69c
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19c2f008e932cbd374a81e20d73eda73d86d1720eafa0f6320a29b7932a61de5
3
  size 4920
vocab.json CHANGED
@@ -1 +1 @@
1
- {"\uc644": 0, "\uc800": 1, "\ub797": 2, "\ucabd": 3, "\uc5c7": 4, "\ubbf8": 5, "\uc7a0": 6, "\ub17c": 7, "\uac9f": 8, "\ucef7": 9, "\uba4b": 10, "\ub611": 11, "\ub534": 12, "\ube60": 13, "\ub189": 14, "\uc090": 15, "\ub36e": 16, "\uafb9": 17, "\ub370": 18, "\ud138": 19, "\uaf3c": 20, "\ub5a0": 21, "\uce61": 22, "\ud480": 23, "\uc70c": 24, "\uac74": 25, "\uc2f6": 26, "\uad75": 27, "\ubb54": 28, "\ud604": 29, "\uc798": 30, "\ube68": 31, "\uae4d": 32, "\uac70": 33, "\ub139": 34, "\ub108": 35, "\ub128": 36, "\ucd0c": 37, "\ub9ce": 38, "\ub518": 39, "\ucda9": 40, "\uc794": 41, "\ub9db": 42, "\uba69": 43, "\ub538": 44, "\uc811": 45, "\ub9c8": 46, "\ud55c": 47, "\uc90d": 48, "\ub1a7": 49, "\ubab0": 50, "\uc601": 51, "\uc5f0": 52, "\ub2f9": 53, "\ub123": 54, "\uc12c": 55, "\uc27d": 56, "\ud615": 57, "\ub04c": 58, "\ucc38": 59, "\ud31c": 60, "\ucd95": 61, "\uc500": 62, "\uc694": 63, "\uadc0": 64, "\ucce5": 65, "\ud504": 66, "\uc5d4": 67, "\uacac": 68, "\uc80a": 69, "\uc78e": 70, "\ud0c1": 71, "\ub2ee": 72, "\ucf54": 73, "\uc5d0": 74, "\uca6c": 75, "\uae08": 76, "\uc5c8": 77, "\ub77d": 78, "\ud6fc": 79, "\ucc44": 80, "\ub798": 81, "\ub4f1": 82, "\uac16": 83, "\ubc84": 84, "\ub728": 85, "\ubd09": 86, "\uafb8": 87, "\ubd14": 88, "\uac90": 89, "\ud5e8": 90, "\uba64": 91, "\uacc4": 92, "\uc4f8": 93, "\uae40": 94, "\uadfc": 95, "\ubcbd": 96, "\ub9bd": 97, "\uc608": 98, "\ub978": 99, "\ub0ad": 100, "\ucc98": 101, "\uc369": 102, "\uc820": 103, "\uba4d": 104, "\uc82f": 105, "\ub461": 106, "\ud2c0": 107, "\ucbe4": 108, "\ubbfc": 109, "\ubc00": 110, "\uc92d": 111, "\uc058": 112, "\uc0c9": 113, "\uac1d": 114, "\uba3c": 115, "\ub0d0": 116, "\ub499": 117, "\uac10": 118, "\ud1a0": 119, "\uaf48": 120, "\uaca1": 121, "\uc9f8": 122, "\uace4": 123, "\ub298": 124, "\ub0b8": 125, "\uac1c": 126, "\uba55": 127, "\ub460": 128, "\uad7d": 129, "\ub35c": 130, "\ub2d0": 131, "\uc464": 132, "\uc88b": 133, "\ub9dd": 134, "\ubb58": 135, "\ud63c": 136, "\ud31f": 137, "\ud5d0": 138, "\uc57d": 139, "\ud0c7": 140, "\ub625": 141, "\uce68": 142, "\ub374": 143, "\uc368": 144, "\ud070": 145, "\ubc29": 146, "\ubc1f": 147, "\uacaa": 148, "\ud310": 149, "\uc78a": 150, "\ub07c": 151, "\uce59": 152, "\ub8e9": 153, "\ud45c": 154, "\ud15d": 155, "\ub0c9": 156, "\uc904": 157, "\uadf8": 158, "\uad70": 159, "\ud5f4": 160, "\uc2eb": 161, "\uc2ac": 162, "\uac13": 163, "\uc8fd": 164, "\uad6c": 165, "\uc815": 166, "\uace1": 167, "\ub364": 168, "\uae68": 169, "\ub8fd": 170, "\uae50": 171, "\ub9b0": 172, "\ubb63": 173, "\ud0c4": 174, "\ub0a0": 175, "\uc5c5": 176, "\ub118": 177, "\uce60": 178, "\uc6b8": 179, "\ub791": 180, "\ubc8b": 181, "\ubcbc": 182, "\ub3cc": 183, "\uac8c": 184, "\uc5d8": 185, "\uacf0": 186, "\ub141": 187, "\ud758": 188, "\ube44": 189, "\ubb38": 190, "\ubcc4": 191, "\uafe9": 192, "\uae5c": 193, "\uad00": 194, "\ud3b8": 195, "\ub0ab": 196, "\ud3c9": 197, "\uce69": 198, "\uc797": 199, "\ud1b5": 200, "\ub290": 201, "\ubb18": 202, "\ud751": 203, "\uad7c": 204, "\ud074": 205, "\uc557": 206, "\uc190": 207, "\uc84b": 208, "\uaf2d": 209, ",": 210, "\uc911": 211, "\ubcfc": 212, "\uccd0": 213, "\uaddc": 214, "\ud314": 215, "\uc0c1": 216, "\uc2ed": 217, "\uc13c": 218, "\uc5bc": 219, "\ubb3b": 220, "\uc2b7": 221, "\uba65": 222, "\uc370": 223, "\ub51c": 224, "\uac78": 225, "\ub8cc": 226, "\ub2f5": 227, "\ub837": 228, "\uc790": 229, "\ub4f8": 230, "\ud321": 231, "\ubc8c": 232, "\uacf5": 233, "\ud0dc": 234, "\uba40": 235, "\ubd10": 236, "\uba38": 237, "\ubcf5": 238, "\ub7ff": 239, "\ub985": 240, "\ubeff": 241, "\ubc18": 242, "\uce78": 243, "\uc2ec": 244, "\ud48d": 245, "\uc18c": 246, "\ub2ff": 247, "\ucfe0": 248, "\uba39": 249, "\ub9b4": 250, "\uc77c": 251, "\ub808": 252, "\ub9bc": 253, "\ucc28": 254, "\ud154": 255, "\ub531": 256, "\uc384": 257, "\ub801": 258, "\ub418": 259, "\uad6d": 260, "\ub5b5": 261, "\ud130": 262, "\ud6a8": 263, "\uae38": 264, "\uc591": 265, "\uc218": 266, "\uc554": 267, "\uc124": 268, "\uc2f8": 269, "\ub7ec": 270, "\ub458": 271, "\ubc88": 272, "\uad74": 273, "\ub110": 274, "\ucd94": 275, "\ubabb": 276, "\uc0bc": 277, "\ub828": 278, "\ub46c": 279, "\ud37c": 280, "\uae5f": 281, "\uc9d1": 282, "\uc0cc": 283, "\uc606": 284, "\ud788": 285, "\uaca0": 286, "\uadf9": 287, "\ub97c": 288, "\uc704": 289, "\uacfc": 290, "\uc6b4": 291, "\ubcc0": 292, "\ube48": 293, "\uce74": 294, "\uc2dc": 295, "\uac00": 296, "\uc584": 297, "\ub550": 298, "\ub4ec": 299, "\ud5d8": 300, "\ub488": 301, "\uaf2c": 302, "\uc788": 303, "\ubaa8": 304, "\ud638": 305, "\uc2f1": 306, "\ub7c9": 307, "\uba67": 308, "\uaed1": 309, "\ub148": 310, "\uc14b": 311, "\ucf00": 312, "\uaf34": 313, "\ud280": 314, "\uc624": 315, "\ube57": 316, "\ud478": 317, "\ubb47": 318, "\uc2f9": 319, "\ubc25": 320, "\uc2a8": 321, "\uba70": 322, "\uc90f": 323, "\uada4": 324, "\uac11": 325, "\ud6cd": 326, "\ub179": 327, "\uc0b4": 328, "\ub4e4": 329, "\uc831": 330, "\ud559": 331, "\ud230": 332, "\uc52c": 333, "\uc838": 334, "\uc796": 335, "\ub150": 336, "\uc2a4": 337, "\ub048": 338, "\uc824": 339, "\ud5f7": 340, "\ucee4": 341, "\ub974": 342, "\ucc45": 343, "\ub151": 344, "\uc4f0": 345, "\ub2cc": 346, "\uc73c": 347, "\uc37b": 348, "\uaef4": 349, "\uce5c": 350, "\ubf51": 351, "\uc131": 352, "\ub0e5": 353, "\ub214": 354, "\uc5b5": 355, "\uc22b": 356, "\uacbd": 357, "\ub4a4": 358, "\ucf20": 359, "\ubabd": 360, "\uad58": 361, "\ub2f4": 362, "\uc881": 363, "\uc81c": 364, "\uae14": 365, "\uc548": 366, "\uc880": 367, "\uc740": 368, "\uccd4": 369, "\ub2ec": 370, "\uc0db": 371, "\ub530": 372, "\uc1a1": 373, "\ub840": 374, "?": 375, "\ub780": 376, "\ub9e4": 377, "\ucd5c": 378, "\uc83c": 379, "\ub807": 380, "\ub465": 381, "\uc538": 382, "\ub3c4": 383, "\uae30": 384, "\uc220": 385, "\ub8e8": 386, "\ub860": 387, "\uccad": 388, "\uc9c0": 389, "\ub0b3": 390, "\ud574": 391, "\uc545": 392, "\uace8": 393, "\ub824": 394, "\ud569": 395, "\uc14d": 396, "\ubfdc": 397, "\ubcd1": 398, "\ub481": 399, "\uc2c0": 400, "\ubb50": 401, "\ub4fc": 402, "\ud30c": 403, "\ub198": 404, "\ub355": 405, "\ub057": 406, "\uc6a9": 407, "\ub354": 408, "\ub178": 409, "\uae34": 410, "\ub5a8": 411, "\ubca1": 412, "\uad7f": 413, "\uc530": 414, "\uc54a": 415, "\ubcb5": 416, "\uc6c3": 417, "\uc870": 418, "\ud22c": 419, "\uba87": 420, "\uad50": 421, "\ud2b8": 422, "\ub871": 423, "\ubd84": 424, "\ub144": 425, "\ub420": 426, "\ub834": 427, "\uba74": 428, "\uc49c": 429, "\ud5f9": 430, "\uaecf": 431, "\ubc97": 432, "\ud68c": 433, "\ub2e8": 434, "\uc758": 435, "\uac19": 436, "\ubfcc": 437, "\uc6d0": 438, "\ucf69": 439, "\ub825": 440, "\ucc29": 441, "\uc9c8": 442, "\uc2e4": 443, "\ub294": 444, "\uc998": 445, "\uc778": 446, "\uacb8": 447, "\uaf07": 448, "\uc934": 449, "\ud2bf": 450, "\ud734": 451, "\uacb0": 452, "\ub839": 453, "\uacbb": 454, "\uccb4": 455, "\ub958": 456, "\ub7b5": 457, "\ubb34": 458, "\ud06c": 459, "\uc094": 460, "\ucef4": 461, "\uce6d": 462, "\uc9e4": 463, "\ub7f0": 464, "\ud544": 465, "\ubaa9": 466, "\ub9c9": 467, "\ub2a0": 468, "\ub9bf": 469, "\ub85d": 470, "\ub358": 471, "\uc544": 472, "\uaca9": 473, "\uc5f4": 474, "\ud655": 475, "\ub2e4": 476, "\ubc15": 477, "\uc9c1": 478, "\uaf3d": 479, "\uc313": 480, "\ub044": 481, "\ud5d9": 482, "\ucb64": 483, "\ub77c": 484, "\ub9cc": 485, "\ub274": 486, "\uc8fc": 487, "\ucc9c": 488, "\uc120": 489, "\ubc16": 490, "\uc5fc": 491, "\uc2b9": 492, "\ud5e4": 493, "\ub098": 494, "\ucc2e": 495, "\ubd05": 496, "\uc0dd": 497, "\ud640": 498, "\ub4b7": 499, "\ucabc": 500, "\uac83": 501, "\ud1f4": 502, "\uccb8": 503, "\ub299": 504, "\uacf1": 505, "\uc37d": 506, "\ub300": 507, "\ubc11": 508, "\ud6c4": 509, "\uba58": 510, "\uaff0": 511, "\ud5c8": 512, "\uacf3": 513, "\uc653": 514, "\uc751": 515, "\uc5e3": 516, "\uc784": 517, "\uca99": 518, "\uc7a5": 519, "\uc871": 520, "\ub194": 521, "\uade4": 522, "\uaed8": 523, "\ub73b": 524, "\uc158": 525, "\ub054": 526, "\ub4e0": 527, "\uc640": 528, "\ubb35": 529, "\uc57c": 530, "\ubc14": 531, "\ubcf4": 532, "\uc11d": 533, "\uac77": 534, "\uc5c6": 535, "\ucc1d": 536, "\ucd10": 537, "\ub81b": 538, "\ud65c": 539, "\ud0a4": 540, "\ub878": 541, "\ub4dc": 542, "\ub2b4": 543, "\ub3c5": 544, "\uc90c": 545, "\ud654": 546, "\uc60c": 547, "\uc918": 548, "\ud2f0": 549, "\uc75c": 550, "\ub5a1": 551, "\uc2dd": 552, "\uc529": 553, "\ubd81": 554, "\ucd9c": 555, "\uc74d": 556, "\ud568": 557, "\ud07c": 558, "\ub9ac": 559, "\uc988": 560, "\uc5c9": 561, "\uac15": 562, "\ub78c": 563, "\ubd90": 564, "\uc628": 565, "\uc300": 566, "\uba54": 567, "\ub2a6": 568, "\ud78c": 569, "\ud56d": 570, "\ud134": 571, "\uc0ac": 572, "\ub124": 573, "\ub369": 574, "\ub193": 575, "\u1161": 576, "\uc2b5": 577, "\uc167": 578, "\ub80c": 579, "\ub1a9": 580, "\uc7ac": 581, "\ub153": 582, "\uac24": 583, "\ubcb3": 584, "\ud1a7": 585, "\ubc24": 586, "\ud6c8": 587, "\uc61b": 588, "\ubd88": 589, "\uc6b1": 590, "\uc874": 591, "\uc6d4": 592, "\ub7fc": 593, "\uc785": 594, "\uc54c": 595, "\ube4c": 596, "\ub490": 597, "\ubed0": 598, "\ucd98": 599, "\u11af": 600, "\uc720": 601, "\ub47f": 602, "\ub1a4": 603, "\uc804": 604, "\uc816": 605, "\uba85": 606, "\ud488": 607, "\ubc85": 608, "\uc11c": 609, "\ud798": 610, "\uc250": 611, "\uc9c7": 612, "\uc55a": 613, "\ud0d1": 614, "\uac85": 615, "\uc721": 616, "\ud611": 617, "\uc74c": 618, "\uc30d": 619, "\uc598": 620, "\uc744": 621, "\uc774": 622, "\ub5bb": 623, "\uc791": 624, ".": 625, "\ub220": 626, "\ubca8": 627, "\uc495": 628, "\uacc1": 629, "\ub9f7": 630, "\ud3ec": 631, "\uc55e": 632, "\uac08": 633, "\ubbff": 634, "\uc5b8": 635, "\uace0": 636, "\ub3d9": 637, "\uafc0": 638, "\ubd25": 639, "\ud2b9": 640, "\ub8d0": 641, "\ucc30": 642, "\uac2f": 643, "\ub9de": 644, "\ub3fc": 645, "\uc7c1": 646, "\ubca0": 647, "\ub09c": 648, "\uae4c": 649, "\ud639": 650, "\ubc27": 651, "\ub78f": 652, "\ubcf8": 653, "\ucb49": 654, "\uc9d3": 655, "\uad82": 656, "\uc9f1": 657, "\ubc94": 658, "\uc787": 659, "\ubc95": 660, "\ud2bc": 661, "\ub20c": 662, "\uc5ec": 663, "\uc62e": 664, "\uc5b4": 665, "\ub984": 666, "\ud53c": 667, "\uac04": 668, "\ubd99": 669, "\ud14c": 670, "\ub9d0": 671, "\ud560": 672, "\ucca0": 673, "\uc138": 674, "\uac07": 675, "\uc5bd": 676, "\u1101": 677, "\uc50c": 678, "\uc6c5": 679, "\uc18d": 680, "\uc528": 681, "\ubc30": 682, "\uc559": 683, "\uc6e8": 684, "\ub2c8": 685, "\ub155": 686, "\ub9e8": 687, "\uc6b0": 688, "\ud750": 689, "\uc560": 690, "\uc5ed": 691, "\ub140": 692, "\uae61": 693, "\uc5ff": 694, "\ub54c": 695, "\ub7fd": 696, "\uc65c": 697, "\ubc1c": 698, "\uc84d": 699, "\ub137": 700, "\uc808": 701, "\ub0a8": 702, "\uca4c": 703, "\uc9d0": 704, "\ub055": 705, "\uad11": 706, "\uc580": 707, "\uc637": 708, "\ub0ae": 709, "\uc775": 710, "\ucf1c": 711, "\uca0b": 712, "\uc5e5": 713, "\uc678": 714, "\ub2d9": 715, "\ubd80": 716, "\uc26c": 717, "\ud5cc": 718, "\ud76c": 719, "\ud0d3": 720, "\ubb3c": 721, "\ub835": 722, "\uae00": 723, "\ub10c": 724, "\ub0bc": 725, "\uac01": 726, "\uc99d": 727, "\ubfd4": 728, "\ub35f": 729, "\ubd04": 730, "\uc2e0": 731, "\uc813": 732, "\ub610": 733, "\ub85c": 734, "\ud754": 735, "\uce6b": 736, "\ud0c0": 737, "\uc12f": 738, "\uac20": 739, "\uc9dd": 740, "\uaca8": 741, "\uae54": 742, "\uc655": 743, "\ub204": 744, "\ub498": 745, "\ub545": 746, "\ubc31": 747, "\ub5a4": 748, "\ub0b4": 749, "\ub385": 750, "\uc9c4": 751, "\ub2d8": 752, "\ub188": 753, "\ub3c8": 754, "\ub48c": 755, "\ub180": 756, "\ub05d": 757, "\ucacc": 758, "\uc21c": 759, "\uc0c8": 760, "\uc6c0": 762, "\uac81": 763, "\uc989": 764, "\ub450": 765, "\ud034": 766, "\uc0b0": 767, "\ud391": 768, "\ucc3e": 769, "\ub4ef": 770, "\ub81d": 771, "\uad81": 772, "\ub72c": 773, "\ucc0d": 774, "\ucc2c": 775, "\ubfd0": 776, "\ub2e5": 777, "\uc900": 778, "\uc885": 779, "\ubbc4": 780, "\uc6c1": 781, "\uac80": 782, "\ubc1b": 783, "\ud150": 784, "\ub18d": 785, "\ud2c8": 786, "\uc5bb": 787, "\uaebc": 788, "\ud558": 789, "\uce58": 790, "\ub514": 791, "\uc62c": 792, "\ucd08": 793, "\ubd23": 794, "\uc801": 795, "\uba83": 796, "\ub7ed": 797, "|": 761, "[UNK]": 798, "[PAD]": 799}
 
1
+ {"\uc5e3": 0, "\uc796": 1, "\uc758": 2, "\uc2e0": 3, "\ubb58": 4, "\uc810": 5, "\ucf1c": 6, "\ub7f0": 7, "\uc560": 8, "\ubb54": 9, "\uc787": 10, "\ub2f4": 11, "\ub9bf": 12, "\ube4c": 13, "\uc300": 14, "\uc9c4": 15, "\ubd80": 16, "\uc54c": 17, "\uaf34": 18, "\uc11d": 19, "\ucca0": 20, "\uba38": 21, "\ub7b5": 22, "\ubc31": 23, "\ub985": 24, "\ubabb": 25, "\uc5f4": 26, "\ub35f": 27, "\ub9ac": 28, "\ub5a4": 29, "\ud074": 30, "\uc6b8": 31, "\ube59": 32, "\ub0bc": 33, "\ud574": 34, "\uacc4": 35, "\ub0b3": 36, "\ucf69": 37, "\ud1b5": 38, "\ub85d": 39, "\ud5d9": 40, "\uacbd": 41, "\uc640": 42, "\ucc9c": 43, "\ub8fd": 44, "\ubd84": 45, "\ud6fc": 46, "\uc219": 47, "\uc838": 48, "\ub9ce": 49, "\uae4c": 50, "\ub530": 51, "\ub0e5": 52, "\uc74d": 53, "\uae61": 54, "\uc2b5": 55, "\uc5fc": 56, "\ud6a8": 57, "\ub611": 58, "\ub05d": 59, "\uaed8": 60, "\uacbb": 61, "\ub9cc": 62, "\ubc1c": 63, "\ucf00": 64, "\uc0b6": 65, "\ucde8": 66, "\ud55c": 67, "\uc138": 68, "\ub2a6": 69, "\ubed0": 70, "\uc0d0": 71, "\ucd08": 72, "\ub204": 73, "\uc149": 74, "\ub514": 75, "\uc194": 76, "\ubfcc": 77, "\uafc0": 78, "\ub2e8": 79, "\ub367": 80, "\uc6ec": 81, "\ud0dc": 82, "\uacac": 83, "\uac07": 84, "\ub807": 85, "\ubabd": 86, "\uc694": 87, "\uc6b0": 88, "\uba39": 89, "\ub9bc": 90, "\ub193": 91, "\ub099": 92, "\uc871": 93, "\ubb35": 94, "\ubcb5": 95, "\uc9d0": 96, "\uc800": 97, "\uc18d": 98, "\ud488": 99, "\uc220": 100, "\ud68c": 101, "\uce58": 102, "\uafb8": 103, "\ud751": 104, "\ub878": 105, "\uc49c": 106, "\ud478": 107, "\ub72c": 108, "\ubf51": 109, "\uc644": 110, "\uc74c": 111, "\uc7a0": 112, "\ub461": 113, "\ub9de": 114, "\uc548": 115, "\ubc84": 116, "\uccc7": 117, "\ub188": 118, "\uac15": 119, "\uc654": 120, "\uccad": 121, "\ub2f9": 122, "\ub9e4": 123, "\ucd9c": 124, "\uc21c": 125, "\ubd23": 126, "\ub4ef": 127, "\uc61b": 128, "\ud6cd": 129, "\uc92b": 130, "\ud5cc": 131, "\ub369": 132, "\ub534": 133, "\uc0db": 134, "\uc7ac": 135, "\ub828": 136, "\ub7ed": 137, "\ub801": 138, "\ub47f": 139, "\uae50": 140, "\uc9c0": 141, "\uba74": 142, "\ucee8": 143, "\uad6c": 144, "\ud589": 145, "\ud314": 146, "\uba67": 147, "\uc529": 148, "\uce61": 149, "\ud130": 150, "\uc5ec": 151, "\ub044": 152, "\ub04a": 153, "\ucda9": 154, "\ud611": 155, "\ud5f9": 156, "\uce6d": 157, "\uc0c1": 158, "\ubd99": 159, "\ud5c8": 160, "\uc820": 161, "\ubc24": 162, "\ucd94": 163, "\uc5f0": 164, "\ubc45": 165, "\uc998": 166, "\ub150": 167, "\ud0b9": 168, "\uc7a6": 169, "\uc22b": 170, "\uac19": 171, "\uac1d": 172, "\ubca1": 173, "\ubbf8": 174, "\uc778": 175, "\uc99d": 176, "\ube44": 177, "\uca4c": 178, "\ubd88": 179, "\ubc8c": 180, "\ucf54": 181, "\uac20": 182, "\ud5e4": 183, "\ubb49": 184, "\ub545": 185, "\ub798": 186, "\uc55a": 187, "\u1101": 188, "\uc2a4": 189, "\uc0cc": 190, "\ub139": 191, "\ub0a0": 192, "\ub81b": 193, "\uccb4": 194, "\uce60": 195, "\uc653": 196, "\ub048": 197, "\uc559": 198, "\ubb34": 199, "\ubb63": 200, "\uaca1": 201, "\ud63c": 202, "\ub098": 203, "\uba4b": 204, "\uc704": 205, "\ub9db": 206, "\ub153": 207, "\uc785": 208, "\ub178": 209, "\uc881": 210, "\uad00": 211, "\ud0d3": 212, "\ub140": 213, "\ubc29": 214, "\ud45c": 215, "\uc90f": 216, "\uc9c8": 217, "\ud2b8": 218, "\uc120": 219, "\ud770": 220, "\ub9dd": 221, "\ud3ec": 222, "\ub3c5": 223, "\uad7c": 224, "\ud798": 225, "\uc9dd": 226, "\uae34": 227, "\uba58": 228, "\uc250": 229, "\ub155": 230, "\ub4f1": 231, "\ub7f4": 232, "\ub458": 233, "\uc90d": 234, "\uc5b4": 235, "\ud558": 236, "\ub791": 237, "\uc9d3": 238, "\ub418": 239, "\uc0b4": 240, "\uae45": 241, "\ubca8": 242, "\ucd5c": 243, "\uaebc": 244, "\uc591": 245, "\uc14d": 246, "\ud2c0": 247, "\uc815": 248, "\uc5d0": 249, "\ubb3b": 250, "\uc0dd": 251, "\ucad9": 252, "\ub180": 253, "\ub7ec": 254, "\ubb3c": 255, "\uac14": 256, ".": 257, "\ud0a4": 258, "\uac1c": 259, "\uc60c": 260, "\ub385": 261, "\uc9e0": 262, "\uc635": 263, "\ud1f4": 264, "\uac80": 265, "\uac74": 266, "\uc900": 267, "\uc2ec": 268, "\ub48c": 269, "\uca6c": 270, "\uc813": 271, "\ub0a9": 272, "\ub108": 273, "\uc57c": 274, "\ub18d": 275, "\ube48": 276, "\ud06c": 277, "\uc62c": 278, "\ud70f": 279, "\uad81": 280, "\uc5c4": 281, "\ud14c": 282, "\ub54c": 283, "\ub290": 284, "\uc2f8": 285, "\uc528": 286, "\uc090": 287, "\ub09c": 288, "\ube68": 289, "\ucb49": 290, "\uc637": 291, "\u11af": 292, "\uc12c": 293, "\uac8c": 294, "\ud654": 295, "\uc8fc": 296, "\ucc38": 297, "\ub4e4": 298, "\uc904": 299, "\ub8d0": 300, "\ubd81": 301, "\ud504": 302, "\uafb9": 303, "\ub35c": 304, "\uc6e0": 305, "\ud78c": 306, "\uc500": 307, "\ub1e8": 308, "\uc918": 309, "\uae08": 310, "\ubcb3": 311, "\uc167": 312, "\ud480": 313, "\ub797": 314, "\ub837": 316, "\uc5e5": 317, "\uc5b8": 318, "\ub825": 319, "\uc5ed": 320, "\ub3d7": 321, "\ubcf8": 322, "\ub144": 323, "\uac78": 324, "\ub17c": 325, "\uc0c9": 326, "\uc6b4": 327, "\ub625": 328, ",": 329, "\uc880": 330, "\ud604": 331, "\uace1": 332, "\ub8cc": 333, "\ucd0c": 334, "\ube57": 335, "\ucd10": 336, "\ub07c": 337, "\ud0d5": 338, "?": 339, "\ub4f8": 340, "\uc0ac": 341, "\uacaa": 342, "\uc9dc": 343, "\ud750": 344, "\ud5a5": 345, "\uba5c": 346, "\uba69": 347, "\uace4": 348, "\uc5c7": 349, "\uc2b9": 350, "\uc788": 351, "\ubd05": 352, "\ub128": 353, "\uc329": 354, "\uc2ef": 355, "\ubd89": 356, "\uac13": 357, "\ub77c": 358, "\ucc1d": 359, "\ud134": 360, "\uc2f9": 361, "\uacf5": 362, "\ub9c9": 363, "\ucc2c": 364, "\uac85": 365, "\ubc25": 366, "\ub2e5": 367, "\uaecd": 368, "\uac24": 369, "\ud65c": 370, "\ud655": 371, "\uadc0": 372, "\ub728": 373, "\ud754": 374, "\ud391": 375, "\ub2e4": 376, "\uc5bc": 377, "\ubc30": 378, "\ucee4": 379, "\ubbfc": 380, "\ub744": 381, "\ubf55": 382, "\ucfe0": 383, "\uc218": 384, "\ucc0d": 385, "\ub3d9": 386, "\uc9c7": 387, "\ubc11": 388, "\ub4e0": 389, "\ub834": 390, "\ubc97": 391, "\ubc1f": 392, "\ud5f7": 393, "\ub7c9": 394, "\uac9f": 395, "\ub550": 396, "\ubb38": 397, "\uc5b5": 398, "\uad82": 399, "\ub0ad": 400, "\uc6b1": 401, "\uacb0": 402, "\uace8": 403, "\ucd1d": 404, "\ubd10": 405, "\uc158": 406, "\ub055": 407, "\ub450": 408, "\ub978": 409, "\uc50c": 410, "\ub36e": 411, "\uad70": 412, "\uc6e8": 413, "\uace7": 414, "\ub9c8": 415, "\uc580": 416, "\ubb47": 417, "\ub6a4": 418, "\ub354": 419, "\uc740": 420, "\ub0d0": 421, "\ub0a8": 422, "\uba55": 423, "\uc598": 424, "\uace0": 425, "\uc608": 426, "\uc601": 427, "\ub2ee": 428, "\ub148": 429, "\uc11c": 430, "\uc384": 431, "\uc2e4": 432, "\ub110": 433, "\uc2ac": 434, "\ub780": 435, "\uc30d": 436, "\uac90": 437, "\uac00": 438, "\ub194": 439, "\ubaa8": 440, "\uad50": 441, "\uc369": 442, "\uc90c": 443, "\ucc3e": 444, "\ub610": 445, "\ucc44": 446, "\uc811": 447, "\ud640": 448, "\uac2f": 449, "\ub3cb": 450, "\ub9d0": 451, "\uc6c1": 452, "\ub3c4": 453, "\uac16": 454, "\uce68": 455, "\ub374": 456, "\ub299": 457, "\ub1a7": 458, "\ud5f4": 459, "\uac08": 460, "\uae14": 461, "\uc54a": 462, "\ud0c1": 463, "\ub2d8": 464, "\ub958": 465, "\ubab0": 466, "\uc624": 467, "\ub0b4": 468, "\uc0ad": 469, "\ub5a0": 470, "\uad6d": 471, "\uad11": 472, "\ubfdc": 473, "\ubd93": 474, "\uae38": 475, "\ub2a0": 476, "\uc797": 477, "\uc5ff": 478, "\uacfc": 479, "\ubc15": 480, "\ub7fd": 481, "\ubbc4": 482, "\ud53c": 483, "\ubd09": 484, "\ud07c": 485, "\uaf48": 486, "\uc6d0": 487, "\uc744": 488, "\ud615": 489, "\uc65c": 490, "\ubcf5": 491, "\ud0c4": 492, "\ub8e9": 493, "\ub4a4": 494, "\ucd95": 495, "\uc2f6": 496, "\uc720": 497, "\ub531": 498, "\ubed8": 499, "\uaca8": 500, "\ub85c": 501, "\ub860": 502, "\ud1a0": 503, "\uae4d": 504, "\uc5bd": 505, "\ub300": 506, "\ud56b": 507, "\ubc14": 508, "\uc790": 509, "\uce74": 510, "\uc554": 511, "\ud230": 512, "\uae30": 513, "\uc370": 514, "\ub1a4": 515, "\ub488": 516, "\ud5e8": 517, "\ub808": 518, "\ud0c0": 519, "\uc5d4": 520, "\uae5f": 521, "\uac10": 522, "\uba40": 523, "\ub538": 524, "\uc625": 525, "\uaf2c": 526, "\uaf3c": 527, "\uafe9": 528, "\ub5a1": 529, "\uad74": 530, "\uac01": 531, "\uaff0": 532, "\uade4": 533, "\uc791": 534, "\uac04": 535, "\uc88b": 536, "\uba4d": 537, "\ub9b0": 538, "\uc0c8": 539, "\ubc00": 540, "\uacf1": 541, "\uc678": 542, "\uc2dd": 543, "\uaca0": 544, "\uc154": 545, "\ub20c": 546, "\uc6c0": 547, "\uc804": 548, "\uc124": 549, "\ub81d": 550, "\ub2eb": 551, "\ub4ec": 552, "\ucabd": 553, "\ubcf4": 554, "\uc0b0": 555, "\uc2a8": 556, "\ud560": 557, "\ub358": 558, "\ucd98": 559, "\uba54": 560, "\ubca0": 561, "\ucf65": 562, "\ub3cc": 563, "\ub365": 564, "\uadf9": 565, "\ucabc": 566, "\uadfc": 567, "\ud5d0": 568, "\ub2ac": 569, "\ub215": 570, "\ubc99": 571, "\uba65": 572, "\uc81c": 573, "\uc131": 574, "\uc368": 575, "\uc2ed": 576, "\uc12f": 577, "\uc4f0": 578, "\ub2f5": 579, "\uc9c1": 580, "\uc655": 581, "\ub057": 582, "\uc557": 583, "\ub984": 584, "\uc606": 585, "\ubfd4": 586, "\uc78a": 587, "\ub8e8": 588, "\ucef4": 589, "\uc6d4": 590, "\ub3c8": 591, "\ud1a7": 592, "\uba85": 593, "\uc190": 594, "\uae54": 595, "\uc7a1": 596, "\ud2bf": 597, "\uc989": 598, "\ud64d": 599, "\uc9d1": 600, "\uc84d": 601, "\ub46c": 602, "\uc808": 603, "\ubc88": 604, "\uc774": 605, "\uac83": 606, "\ud0d1": 607, "\uc824": 608, "\uadf8": 609, "\uad7f": 610, "\uc5c9": 611, "\uc870": 612, "\ucc28": 613, "\ub0ab": 614, "\uc7a5": 615, "\uc6a9": 616, "\ub9b4": 617, "\ub123": 618, "\uc530": 619, "\ub7ff": 620, "\u1161": 621, "\uae00": 622, "\uc628": 623, "\uc2dc": 624, "\uccd0": 625, "\ud37c": 626, "\uce6b": 627, "\ub0c9": 628, "\uc5c5": 629, "\ud31c": 630, "\ubcbd": 631, "\ub824": 632, "\ud30c": 633, "\ud5d8": 634, "\uc878": 635, "\uc4f8": 636, "\ubcc4": 637, "\ud034": 638, "\uc885": 639, "\ub2c8": 640, "\uac77": 641, "\ud2c8": 642, "\ud31f": 643, "\ub77d": 644, "\ud15f": 645, "\ub839": 646, "\ucc48": 647, "\ubc95": 648, "\ub2ec": 649, "\ub871": 650, "\uc13c": 651, "\ub118": 652, "\ud48d": 653, "\ubc27": 654, "\ub490": 655, "\uc8fd": 656, "\uce78": 657, "\ucbe4": 658, "\ucce5": 659, "\ub124": 660, "\uc544": 661, "\ubd25": 662, "\uc831": 663, "\uc0bc": 664, "\ub518": 665, "\ub78c": 666, "\ubc18": 667, "\uac70": 668, "\ubc16": 669, "\uc84b": 670, "\uc55e": 671, "\uaf5d": 672, "\ud6c4": 673, "\uc6c3": 674, "\ub7fc": 675, "\uc77c": 676, "\uc7c1": 677, "\ubeff": 678, "\uacf0": 679, "\ub304": 680, "\ub4dc": 681, "\ub974": 682, "\ud638": 683, "\uaf2d": 684, "\uc2f1": 685, "\uc721": 686, "\ubcc0": 687, "\ub95c": 688, "\ubd14": 689, "\ub9bd": 690, "\ub1a9": 691, "\ud788": 692, "\ud76c": 693, "\ub294": 694, "\ub465": 695, "\ucacc": 696, "\uc5c6": 697, "\ub054": 698, "\ud138": 699, "\ubc1b": 700, "\uaed1": 701, "\uc911": 702, "\uc37b": 703, "\ubd04": 704, "\ub04c": 705, "\ub840": 706, "\uc798": 707, "\ud321": 708, "\uc695": 709, "\ub0b8": 710, "\uc094": 711, "\ubaa9": 712, "\ucc3d": 713, "\ub80c": 714, "\ud2bc": 715, "\ube60": 716, "\uc22d": 717, "\uc18c": 718, "\ubc8b": 719, "\ub729": 720, "\ucf04": 721, "\ub5b5": 722, "\uacf3": 723, "\ub97c": 724, "\uc775": 725, "\ud639": 726, "\uc801": 727, "\uc058": 728, "\uc6cc": 729, "\ucc98": 730, "\ub481": 731, "\ub370": 732, "\ub179": 733, "\uce7c": 734, "\uc73c": 735, "\ud070": 736, "\ucc14": 737, "\uc874": 738, "\ud2f0": 739, "\ubb50": 740, "\uc62e": 741, "\ub460": 742, "\ud559": 743, "|": 315, "[UNK]": 744, "[PAD]": 745}
wandb/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log CHANGED
@@ -1,32 +1,32 @@
1
- 2024-05-25 17:50:10,434 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.6
2
- 2024-05-25 17:50:10,434 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
- 2024-05-25 17:50:10,434 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-05-25 17:50:10,434 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-05-25 17:50:10,434 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-05-25 17:50:10,434 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
7
- 2024-05-25 17:50:10,434 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
8
- 2024-05-25 17:50:10,434 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
- 2024-05-25 17:50:10,434 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
10
- 2024-05-25 17:50:10,434 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging user logs to /kaggle/working/wandb/run-20240525_175010-8ah63pdc/logs/debug.log
11
- 2024-05-25 17:50:10,434 INFO MainThread:34 [wandb_init.py:_log_setup():522] Logging internal logs to /kaggle/working/wandb/run-20240525_175010-8ah63pdc/logs/debug-internal.log
12
- 2024-05-25 17:50:10,434 INFO MainThread:34 [wandb_init.py:_jupyter_setup():467] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7fbe99ebfe20>
13
- 2024-05-25 17:50:10,435 INFO MainThread:34 [wandb_init.py:init():561] calling init triggers
14
- 2024-05-25 17:50:10,435 INFO MainThread:34 [wandb_init.py:init():568] wandb.init called with sweep_config: {}
15
  config: {}
16
- 2024-05-25 17:50:10,435 INFO MainThread:34 [wandb_init.py:init():611] starting backend
17
- 2024-05-25 17:50:10,435 INFO MainThread:34 [wandb_init.py:init():615] setting up manager
18
- 2024-05-25 17:50:10,437 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
- 2024-05-25 17:50:10,439 INFO MainThread:34 [wandb_init.py:init():623] backend started and connected
20
- 2024-05-25 17:50:10,450 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1299] probe notebook
21
- 2024-05-25 17:50:11,144 INFO MainThread:34 [wandb_init.py:init():715] updated telemetry
22
- 2024-05-25 17:50:11,147 INFO MainThread:34 [wandb_init.py:init():748] communicating run to backend with 90.0 second timeout
23
- 2024-05-25 17:50:11,381 INFO MainThread:34 [wandb_run.py:_on_init():2357] communicating current version
24
- 2024-05-25 17:50:11,442 INFO MainThread:34 [wandb_run.py:_on_init():2366] got version response upgrade_message: "wandb version 0.17.0 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
 
26
- 2024-05-25 17:50:11,444 INFO MainThread:34 [wandb_init.py:init():799] starting run threads in backend
27
- 2024-05-25 17:50:27,456 INFO MainThread:34 [wandb_run.py:_console_start():2335] atexit reg
28
- 2024-05-25 17:50:27,456 INFO MainThread:34 [wandb_run.py:_redirect():2190] redirect: wrap_raw
29
- 2024-05-25 17:50:27,456 INFO MainThread:34 [wandb_run.py:_redirect():2255] Wrapping output streams.
30
- 2024-05-25 17:50:27,456 INFO MainThread:34 [wandb_run.py:_redirect():2280] Redirects installed.
31
- 2024-05-25 17:50:27,457 INFO MainThread:34 [wandb_init.py:init():842] run started, returning control to user process
32
- 2024-05-25 17:50:27,464 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 100000000000000, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May25_17-49-41_8f1fad5fe1d2', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
 
1
+ 2024-05-26 14:13:05,023 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.6
2
+ 2024-05-26 14:13:05,023 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
7
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
8
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
10
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging user logs to /kaggle/working/wandb/run-20240526_141305-wfyhopk0/logs/debug.log
11
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_init.py:_log_setup():522] Logging internal logs to /kaggle/working/wandb/run-20240526_141305-wfyhopk0/logs/debug-internal.log
12
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_init.py:_jupyter_setup():467] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7bfadacb7c10>
13
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_init.py:init():561] calling init triggers
14
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_init.py:init():568] wandb.init called with sweep_config: {}
15
  config: {}
16
+ 2024-05-26 14:13:05,025 INFO MainThread:34 [wandb_init.py:init():611] starting backend
17
+ 2024-05-26 14:13:05,025 INFO MainThread:34 [wandb_init.py:init():615] setting up manager
18
+ 2024-05-26 14:13:05,026 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-05-26 14:13:05,028 INFO MainThread:34 [wandb_init.py:init():623] backend started and connected
20
+ 2024-05-26 14:13:05,040 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1299] probe notebook
21
+ 2024-05-26 14:13:06,043 INFO MainThread:34 [wandb_init.py:init():715] updated telemetry
22
+ 2024-05-26 14:13:06,046 INFO MainThread:34 [wandb_init.py:init():748] communicating run to backend with 90.0 second timeout
23
+ 2024-05-26 14:13:06,476 INFO MainThread:34 [wandb_run.py:_on_init():2357] communicating current version
24
+ 2024-05-26 14:13:06,526 INFO MainThread:34 [wandb_run.py:_on_init():2366] got version response upgrade_message: "wandb version 0.17.0 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
 
26
+ 2024-05-26 14:13:06,527 INFO MainThread:34 [wandb_init.py:init():799] starting run threads in backend
27
+ 2024-05-26 14:13:22,559 INFO MainThread:34 [wandb_run.py:_console_start():2335] atexit reg
28
+ 2024-05-26 14:13:22,559 INFO MainThread:34 [wandb_run.py:_redirect():2190] redirect: wrap_raw
29
+ 2024-05-26 14:13:22,559 INFO MainThread:34 [wandb_run.py:_redirect():2255] Wrapping output streams.
30
+ 2024-05-26 14:13:22,559 INFO MainThread:34 [wandb_run.py:_redirect():2280] Redirects installed.
31
+ 2024-05-26 14:13:22,560 INFO MainThread:34 [wandb_init.py:init():842] run started, returning control to user process
32
+ 2024-05-26 14:13:22,567 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 100000000000000, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May26_14-12-33_142d81d78b72', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240526_141305-wfyhopk0/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240526_141305-wfyhopk0/files/config.yaml ADDED
@@ -0,0 +1,846 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.6
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1716732785.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 105
27
+ 2:
28
+ - 1
29
+ - 2
30
+ - 3
31
+ - 5
32
+ - 11
33
+ - 12
34
+ - 49
35
+ - 51
36
+ - 53
37
+ - 55
38
+ - 71
39
+ - 105
40
+ 3:
41
+ - 7
42
+ - 23
43
+ 4: 3.10.13
44
+ 5: 0.16.6
45
+ 6: 4.39.3
46
+ 8:
47
+ - 1
48
+ - 2
49
+ - 5
50
+ 9:
51
+ 1: transformers_trainer
52
+ 13: linux-x86_64
53
+ m:
54
+ - 1: train/global_step
55
+ 6:
56
+ - 3
57
+ - 1: train/loss
58
+ 5: 1
59
+ 6:
60
+ - 1
61
+ - 1: train/grad_norm
62
+ 5: 1
63
+ 6:
64
+ - 1
65
+ - 1: train/learning_rate
66
+ 5: 1
67
+ 6:
68
+ - 1
69
+ - 1: train/epoch
70
+ 5: 1
71
+ 6:
72
+ - 1
73
+ return_dict:
74
+ desc: null
75
+ value: true
76
+ output_hidden_states:
77
+ desc: null
78
+ value: false
79
+ output_attentions:
80
+ desc: null
81
+ value: false
82
+ torchscript:
83
+ desc: null
84
+ value: false
85
+ torch_dtype:
86
+ desc: null
87
+ value: null
88
+ use_bfloat16:
89
+ desc: null
90
+ value: false
91
+ tf_legacy_loss:
92
+ desc: null
93
+ value: false
94
+ pruned_heads:
95
+ desc: null
96
+ value: {}
97
+ tie_word_embeddings:
98
+ desc: null
99
+ value: true
100
+ chunk_size_feed_forward:
101
+ desc: null
102
+ value: 0
103
+ is_encoder_decoder:
104
+ desc: null
105
+ value: false
106
+ is_decoder:
107
+ desc: null
108
+ value: false
109
+ cross_attention_hidden_size:
110
+ desc: null
111
+ value: null
112
+ add_cross_attention:
113
+ desc: null
114
+ value: false
115
+ tie_encoder_decoder:
116
+ desc: null
117
+ value: false
118
+ max_length:
119
+ desc: null
120
+ value: 20
121
+ min_length:
122
+ desc: null
123
+ value: 0
124
+ do_sample:
125
+ desc: null
126
+ value: false
127
+ early_stopping:
128
+ desc: null
129
+ value: false
130
+ num_beams:
131
+ desc: null
132
+ value: 1
133
+ num_beam_groups:
134
+ desc: null
135
+ value: 1
136
+ diversity_penalty:
137
+ desc: null
138
+ value: 0.0
139
+ temperature:
140
+ desc: null
141
+ value: 1.0
142
+ top_k:
143
+ desc: null
144
+ value: 50
145
+ top_p:
146
+ desc: null
147
+ value: 1.0
148
+ typical_p:
149
+ desc: null
150
+ value: 1.0
151
+ repetition_penalty:
152
+ desc: null
153
+ value: 1.0
154
+ length_penalty:
155
+ desc: null
156
+ value: 1.0
157
+ no_repeat_ngram_size:
158
+ desc: null
159
+ value: 0
160
+ encoder_no_repeat_ngram_size:
161
+ desc: null
162
+ value: 0
163
+ bad_words_ids:
164
+ desc: null
165
+ value: null
166
+ num_return_sequences:
167
+ desc: null
168
+ value: 1
169
+ output_scores:
170
+ desc: null
171
+ value: false
172
+ return_dict_in_generate:
173
+ desc: null
174
+ value: false
175
+ forced_bos_token_id:
176
+ desc: null
177
+ value: null
178
+ forced_eos_token_id:
179
+ desc: null
180
+ value: null
181
+ remove_invalid_values:
182
+ desc: null
183
+ value: false
184
+ exponential_decay_length_penalty:
185
+ desc: null
186
+ value: null
187
+ suppress_tokens:
188
+ desc: null
189
+ value: null
190
+ begin_suppress_tokens:
191
+ desc: null
192
+ value: null
193
+ architectures:
194
+ desc: null
195
+ value:
196
+ - Wav2Vec2ForPreTraining
197
+ finetuning_task:
198
+ desc: null
199
+ value: null
200
+ id2label:
201
+ desc: null
202
+ value:
203
+ '0': LABEL_0
204
+ '1': LABEL_1
205
+ label2id:
206
+ desc: null
207
+ value:
208
+ LABEL_0: 0
209
+ LABEL_1: 1
210
+ tokenizer_class:
211
+ desc: null
212
+ value: null
213
+ prefix:
214
+ desc: null
215
+ value: null
216
+ bos_token_id:
217
+ desc: null
218
+ value: 1
219
+ pad_token_id:
220
+ desc: null
221
+ value: 0
222
+ eos_token_id:
223
+ desc: null
224
+ value: 2
225
+ sep_token_id:
226
+ desc: null
227
+ value: null
228
+ decoder_start_token_id:
229
+ desc: null
230
+ value: null
231
+ task_specific_params:
232
+ desc: null
233
+ value: null
234
+ problem_type:
235
+ desc: null
236
+ value: null
237
+ _name_or_path:
238
+ desc: null
239
+ value: facebook/wav2vec2-base
240
+ transformers_version:
241
+ desc: null
242
+ value: 4.39.3
243
+ freeze_feat_extract_train:
244
+ desc: null
245
+ value: true
246
+ mask_channel_length:
247
+ desc: null
248
+ value: 10
249
+ mask_channel_min_space:
250
+ desc: null
251
+ value: 1
252
+ mask_channel_other:
253
+ desc: null
254
+ value: 0.0
255
+ mask_channel_prob:
256
+ desc: null
257
+ value: 0.0
258
+ mask_channel_selection:
259
+ desc: null
260
+ value: static
261
+ mask_time_min_space:
262
+ desc: null
263
+ value: 1
264
+ mask_time_other:
265
+ desc: null
266
+ value: 0.0
267
+ mask_time_selection:
268
+ desc: null
269
+ value: static
270
+ model_type:
271
+ desc: null
272
+ value: wav2vec2
273
+ no_mask_channel_overlap:
274
+ desc: null
275
+ value: false
276
+ no_mask_time_overlap:
277
+ desc: null
278
+ value: false
279
+ num_feat_extract_layers:
280
+ desc: null
281
+ value: 7
282
+ hidden_size:
283
+ desc: null
284
+ value: 768
285
+ feat_extract_norm:
286
+ desc: null
287
+ value: group
288
+ feat_extract_activation:
289
+ desc: null
290
+ value: gelu
291
+ conv_dim:
292
+ desc: null
293
+ value:
294
+ - 512
295
+ - 512
296
+ - 512
297
+ - 512
298
+ - 512
299
+ - 512
300
+ - 512
301
+ conv_stride:
302
+ desc: null
303
+ value:
304
+ - 5
305
+ - 2
306
+ - 2
307
+ - 2
308
+ - 2
309
+ - 2
310
+ - 2
311
+ conv_kernel:
312
+ desc: null
313
+ value:
314
+ - 10
315
+ - 3
316
+ - 3
317
+ - 3
318
+ - 3
319
+ - 2
320
+ - 2
321
+ conv_bias:
322
+ desc: null
323
+ value: false
324
+ num_conv_pos_embeddings:
325
+ desc: null
326
+ value: 128
327
+ num_conv_pos_embedding_groups:
328
+ desc: null
329
+ value: 16
330
+ num_hidden_layers:
331
+ desc: null
332
+ value: 12
333
+ intermediate_size:
334
+ desc: null
335
+ value: 3072
336
+ hidden_act:
337
+ desc: null
338
+ value: gelu
339
+ num_attention_heads:
340
+ desc: null
341
+ value: 12
342
+ hidden_dropout:
343
+ desc: null
344
+ value: 0.1
345
+ attention_dropout:
346
+ desc: null
347
+ value: 0.1
348
+ activation_dropout:
349
+ desc: null
350
+ value: 0.0
351
+ feat_proj_dropout:
352
+ desc: null
353
+ value: 0.1
354
+ final_dropout:
355
+ desc: null
356
+ value: 0.0
357
+ layerdrop:
358
+ desc: null
359
+ value: 0.0
360
+ layer_norm_eps:
361
+ desc: null
362
+ value: 1.0e-05
363
+ initializer_range:
364
+ desc: null
365
+ value: 0.02
366
+ vocab_size:
367
+ desc: null
368
+ value: 100000000000000
369
+ do_stable_layer_norm:
370
+ desc: null
371
+ value: false
372
+ use_weighted_layer_sum:
373
+ desc: null
374
+ value: false
375
+ apply_spec_augment:
376
+ desc: null
377
+ value: true
378
+ mask_time_prob:
379
+ desc: null
380
+ value: 0.05
381
+ mask_time_length:
382
+ desc: null
383
+ value: 10
384
+ mask_time_min_masks:
385
+ desc: null
386
+ value: 2
387
+ mask_feature_prob:
388
+ desc: null
389
+ value: 0.0
390
+ mask_feature_length:
391
+ desc: null
392
+ value: 10
393
+ mask_feature_min_masks:
394
+ desc: null
395
+ value: 0
396
+ num_codevectors_per_group:
397
+ desc: null
398
+ value: 320
399
+ num_codevector_groups:
400
+ desc: null
401
+ value: 2
402
+ contrastive_logits_temperature:
403
+ desc: null
404
+ value: 0.1
405
+ feat_quantizer_dropout:
406
+ desc: null
407
+ value: 0.0
408
+ num_negatives:
409
+ desc: null
410
+ value: 100
411
+ codevector_dim:
412
+ desc: null
413
+ value: 256
414
+ proj_codevector_dim:
415
+ desc: null
416
+ value: 256
417
+ diversity_loss_weight:
418
+ desc: null
419
+ value: 0.1
420
+ ctc_loss_reduction:
421
+ desc: null
422
+ value: sum
423
+ ctc_zero_infinity:
424
+ desc: null
425
+ value: false
426
+ add_adapter:
427
+ desc: null
428
+ value: false
429
+ adapter_kernel_size:
430
+ desc: null
431
+ value: 3
432
+ adapter_stride:
433
+ desc: null
434
+ value: 2
435
+ num_adapter_layers:
436
+ desc: null
437
+ value: 3
438
+ output_hidden_size:
439
+ desc: null
440
+ value: 768
441
+ adapter_attn_dim:
442
+ desc: null
443
+ value: null
444
+ classifier_proj_size:
445
+ desc: null
446
+ value: 256
447
+ tdnn_dim:
448
+ desc: null
449
+ value:
450
+ - 512
451
+ - 512
452
+ - 512
453
+ - 512
454
+ - 1500
455
+ tdnn_kernel:
456
+ desc: null
457
+ value:
458
+ - 5
459
+ - 3
460
+ - 3
461
+ - 1
462
+ - 1
463
+ tdnn_dilation:
464
+ desc: null
465
+ value:
466
+ - 1
467
+ - 2
468
+ - 3
469
+ - 1
470
+ - 1
471
+ xvector_output_dim:
472
+ desc: null
473
+ value: 512
474
+ output_dir:
475
+ desc: null
476
+ value: /kaggle/working/
477
+ overwrite_output_dir:
478
+ desc: null
479
+ value: false
480
+ do_train:
481
+ desc: null
482
+ value: false
483
+ do_eval:
484
+ desc: null
485
+ value: true
486
+ do_predict:
487
+ desc: null
488
+ value: false
489
+ evaluation_strategy:
490
+ desc: null
491
+ value: steps
492
+ prediction_loss_only:
493
+ desc: null
494
+ value: false
495
+ per_device_train_batch_size:
496
+ desc: null
497
+ value: 8
498
+ per_device_eval_batch_size:
499
+ desc: null
500
+ value: 8
501
+ per_gpu_train_batch_size:
502
+ desc: null
503
+ value: null
504
+ per_gpu_eval_batch_size:
505
+ desc: null
506
+ value: null
507
+ gradient_accumulation_steps:
508
+ desc: null
509
+ value: 1
510
+ eval_accumulation_steps:
511
+ desc: null
512
+ value: null
513
+ eval_delay:
514
+ desc: null
515
+ value: 0
516
+ learning_rate:
517
+ desc: null
518
+ value: 0.0001
519
+ weight_decay:
520
+ desc: null
521
+ value: 0.005
522
+ adam_beta1:
523
+ desc: null
524
+ value: 0.9
525
+ adam_beta2:
526
+ desc: null
527
+ value: 0.999
528
+ adam_epsilon:
529
+ desc: null
530
+ value: 1.0e-08
531
+ max_grad_norm:
532
+ desc: null
533
+ value: 1.0
534
+ num_train_epochs:
535
+ desc: null
536
+ value: 30
537
+ max_steps:
538
+ desc: null
539
+ value: -1
540
+ lr_scheduler_type:
541
+ desc: null
542
+ value: linear
543
+ lr_scheduler_kwargs:
544
+ desc: null
545
+ value: {}
546
+ warmup_ratio:
547
+ desc: null
548
+ value: 0.0
549
+ warmup_steps:
550
+ desc: null
551
+ value: 1000
552
+ log_level:
553
+ desc: null
554
+ value: passive
555
+ log_level_replica:
556
+ desc: null
557
+ value: warning
558
+ log_on_each_node:
559
+ desc: null
560
+ value: true
561
+ logging_dir:
562
+ desc: null
563
+ value: /kaggle/working/runs/May26_14-12-33_142d81d78b72
564
+ logging_strategy:
565
+ desc: null
566
+ value: steps
567
+ logging_first_step:
568
+ desc: null
569
+ value: false
570
+ logging_steps:
571
+ desc: null
572
+ value: 500
573
+ logging_nan_inf_filter:
574
+ desc: null
575
+ value: true
576
+ save_strategy:
577
+ desc: null
578
+ value: steps
579
+ save_steps:
580
+ desc: null
581
+ value: 500
582
+ save_total_limit:
583
+ desc: null
584
+ value: 2
585
+ save_safetensors:
586
+ desc: null
587
+ value: true
588
+ save_on_each_node:
589
+ desc: null
590
+ value: false
591
+ save_only_model:
592
+ desc: null
593
+ value: false
594
+ no_cuda:
595
+ desc: null
596
+ value: false
597
+ use_cpu:
598
+ desc: null
599
+ value: false
600
+ use_mps_device:
601
+ desc: null
602
+ value: false
603
+ seed:
604
+ desc: null
605
+ value: 42
606
+ data_seed:
607
+ desc: null
608
+ value: null
609
+ jit_mode_eval:
610
+ desc: null
611
+ value: false
612
+ use_ipex:
613
+ desc: null
614
+ value: false
615
+ bf16:
616
+ desc: null
617
+ value: false
618
+ fp16:
619
+ desc: null
620
+ value: true
621
+ fp16_opt_level:
622
+ desc: null
623
+ value: O1
624
+ half_precision_backend:
625
+ desc: null
626
+ value: auto
627
+ bf16_full_eval:
628
+ desc: null
629
+ value: false
630
+ fp16_full_eval:
631
+ desc: null
632
+ value: false
633
+ tf32:
634
+ desc: null
635
+ value: null
636
+ local_rank:
637
+ desc: null
638
+ value: 0
639
+ ddp_backend:
640
+ desc: null
641
+ value: null
642
+ tpu_num_cores:
643
+ desc: null
644
+ value: null
645
+ tpu_metrics_debug:
646
+ desc: null
647
+ value: false
648
+ debug:
649
+ desc: null
650
+ value: []
651
+ dataloader_drop_last:
652
+ desc: null
653
+ value: false
654
+ eval_steps:
655
+ desc: null
656
+ value: 500
657
+ dataloader_num_workers:
658
+ desc: null
659
+ value: 0
660
+ dataloader_prefetch_factor:
661
+ desc: null
662
+ value: null
663
+ past_index:
664
+ desc: null
665
+ value: -1
666
+ run_name:
667
+ desc: null
668
+ value: /kaggle/working/
669
+ disable_tqdm:
670
+ desc: null
671
+ value: false
672
+ remove_unused_columns:
673
+ desc: null
674
+ value: true
675
+ label_names:
676
+ desc: null
677
+ value: null
678
+ load_best_model_at_end:
679
+ desc: null
680
+ value: false
681
+ metric_for_best_model:
682
+ desc: null
683
+ value: null
684
+ greater_is_better:
685
+ desc: null
686
+ value: null
687
+ ignore_data_skip:
688
+ desc: null
689
+ value: false
690
+ fsdp:
691
+ desc: null
692
+ value: []
693
+ fsdp_min_num_params:
694
+ desc: null
695
+ value: 0
696
+ fsdp_config:
697
+ desc: null
698
+ value:
699
+ min_num_params: 0
700
+ xla: false
701
+ xla_fsdp_v2: false
702
+ xla_fsdp_grad_ckpt: false
703
+ fsdp_transformer_layer_cls_to_wrap:
704
+ desc: null
705
+ value: null
706
+ accelerator_config:
707
+ desc: null
708
+ value:
709
+ split_batches: false
710
+ dispatch_batches: null
711
+ even_batches: true
712
+ use_seedable_sampler: true
713
+ deepspeed:
714
+ desc: null
715
+ value: null
716
+ label_smoothing_factor:
717
+ desc: null
718
+ value: 0.0
719
+ optim:
720
+ desc: null
721
+ value: adamw_torch
722
+ optim_args:
723
+ desc: null
724
+ value: null
725
+ adafactor:
726
+ desc: null
727
+ value: false
728
+ group_by_length:
729
+ desc: null
730
+ value: true
731
+ length_column_name:
732
+ desc: null
733
+ value: length
734
+ report_to:
735
+ desc: null
736
+ value:
737
+ - tensorboard
738
+ - wandb
739
+ ddp_find_unused_parameters:
740
+ desc: null
741
+ value: null
742
+ ddp_bucket_cap_mb:
743
+ desc: null
744
+ value: null
745
+ ddp_broadcast_buffers:
746
+ desc: null
747
+ value: null
748
+ dataloader_pin_memory:
749
+ desc: null
750
+ value: true
751
+ dataloader_persistent_workers:
752
+ desc: null
753
+ value: false
754
+ skip_memory_metrics:
755
+ desc: null
756
+ value: true
757
+ use_legacy_prediction_loop:
758
+ desc: null
759
+ value: false
760
+ push_to_hub:
761
+ desc: null
762
+ value: true
763
+ resume_from_checkpoint:
764
+ desc: null
765
+ value: null
766
+ hub_model_id:
767
+ desc: null
768
+ value: null
769
+ hub_strategy:
770
+ desc: null
771
+ value: every_save
772
+ hub_token:
773
+ desc: null
774
+ value: <HUB_TOKEN>
775
+ hub_private_repo:
776
+ desc: null
777
+ value: false
778
+ hub_always_push:
779
+ desc: null
780
+ value: false
781
+ gradient_checkpointing:
782
+ desc: null
783
+ value: true
784
+ gradient_checkpointing_kwargs:
785
+ desc: null
786
+ value: null
787
+ include_inputs_for_metrics:
788
+ desc: null
789
+ value: false
790
+ fp16_backend:
791
+ desc: null
792
+ value: auto
793
+ push_to_hub_model_id:
794
+ desc: null
795
+ value: null
796
+ push_to_hub_organization:
797
+ desc: null
798
+ value: null
799
+ push_to_hub_token:
800
+ desc: null
801
+ value: <PUSH_TO_HUB_TOKEN>
802
+ mp_parameters:
803
+ desc: null
804
+ value: ''
805
+ auto_find_batch_size:
806
+ desc: null
807
+ value: false
808
+ full_determinism:
809
+ desc: null
810
+ value: false
811
+ torchdynamo:
812
+ desc: null
813
+ value: null
814
+ ray_scope:
815
+ desc: null
816
+ value: last
817
+ ddp_timeout:
818
+ desc: null
819
+ value: 1800
820
+ torch_compile:
821
+ desc: null
822
+ value: false
823
+ torch_compile_backend:
824
+ desc: null
825
+ value: null
826
+ torch_compile_mode:
827
+ desc: null
828
+ value: null
829
+ dispatch_batches:
830
+ desc: null
831
+ value: null
832
+ split_batches:
833
+ desc: null
834
+ value: null
835
+ include_tokens_per_second:
836
+ desc: null
837
+ value: false
838
+ include_num_input_tokens_seen:
839
+ desc: null
840
+ value: false
841
+ neftune_noise_alpha:
842
+ desc: null
843
+ value: null
844
+ optim_target_modules:
845
+ desc: null
846
+ value: null
wandb/run-20240526_141305-wfyhopk0/files/output.log ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:156: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
2
+ warnings.warn(
3
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
4
+ warnings.warn(
5
+ /opt/conda/lib/python3.10/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:156: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
6
+ warnings.warn(
7
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
wandb/run-20240526_141305-wfyhopk0/files/requirements.txt ADDED
@@ -0,0 +1,862 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.23.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.3
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.3
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.29.3
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.3
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.18.0
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.15.2.45.49
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ blake3==0.2.1
86
+ bleach==6.1.0
87
+ blessed==1.20.0
88
+ blinker==1.7.0
89
+ blis==0.7.10
90
+ blosc2==2.6.2
91
+ bokeh==3.4.1
92
+ boltons==23.1.1
93
+ boto3==1.26.100
94
+ botocore==1.34.69
95
+ bq_helper==0.4.1
96
+ bqplot==0.12.43
97
+ branca==0.7.1
98
+ brewer2mpl==1.4.1
99
+ brotlipy==0.7.0
100
+ cached-property==1.5.2
101
+ cachetools==4.2.4
102
+ cachetools==5.3.2
103
+ catalogue==2.0.10
104
+ catalyst==22.4
105
+ catboost==1.2.3
106
+ category-encoders==2.6.3
107
+ certifi==2024.2.2
108
+ cesium==0.12.1
109
+ cffi==1.16.0
110
+ charset-normalizer==3.3.2
111
+ chex==0.1.86
112
+ cleverhans==4.0.0
113
+ click-plugins==1.1.1
114
+ click==8.1.7
115
+ cligj==0.7.2
116
+ cloud-tpu-client==0.10
117
+ cloud-tpu-profiler==2.4.0
118
+ cloudpathlib==0.16.0
119
+ cloudpickle==2.2.1
120
+ cloudpickle==3.0.0
121
+ cmdstanpy==1.2.2
122
+ colorama==0.4.6
123
+ colorcet==3.1.0
124
+ colorful==0.5.6
125
+ colorlog==6.8.2
126
+ colorlover==0.3.0
127
+ comm==0.2.1
128
+ conda-libmamba-solver==23.7.0
129
+ conda-package-handling==2.2.0
130
+ conda==23.7.4
131
+ conda_package_streaming==0.9.0
132
+ confection==0.1.4
133
+ contextily==1.6.0
134
+ contourpy==1.2.0
135
+ contourpy==1.2.1
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.3.0
148
+ daal==2024.3.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.11
153
+ dask==2024.4.1
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.18.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==7.0.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.8
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.21.1
179
+ earthengine-api==0.1.399
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.51.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.2
220
+ frozenlist==1.4.1
221
+ fsspec==2024.2.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.12.0
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.6.2
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.126.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.5.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.11.0
331
+ ipyvuetify==1.9.4
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.2.0
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jiwer==3.0.4
346
+ jmespath==1.0.1
347
+ joblib==1.4.0
348
+ json5==0.9.14
349
+ jsonpatch==1.33
350
+ jsonpointer==2.4
351
+ jsonschema-specifications==2023.12.1
352
+ jsonschema==4.20.0
353
+ jupyter-console==6.6.3
354
+ jupyter-events==0.9.0
355
+ jupyter-http-over-ws==0.0.8
356
+ jupyter-lsp==1.5.1
357
+ jupyter-server-mathjax==0.2.6
358
+ jupyter-ydoc==0.2.5
359
+ jupyter_client==7.4.9
360
+ jupyter_client==8.6.0
361
+ jupyter_core==5.7.1
362
+ jupyter_server==2.12.5
363
+ jupyter_server_fileid==0.9.1
364
+ jupyter_server_proxy==4.1.0
365
+ jupyter_server_terminals==0.5.1
366
+ jupyter_server_ydoc==0.8.0
367
+ jupyterlab-lsp==5.1.0
368
+ jupyterlab-widgets==3.0.9
369
+ jupyterlab==4.1.6
370
+ jupyterlab_git==0.44.0
371
+ jupyterlab_pygments==0.3.0
372
+ jupyterlab_server==2.25.2
373
+ jupytext==1.16.0
374
+ kaggle-environments==1.14.3
375
+ kaggle==1.6.12
376
+ kagglehub==0.2.3
377
+ keras-cv==0.8.2
378
+ keras-nlp==0.9.3
379
+ keras-tuner==1.4.6
380
+ keras==3.2.1
381
+ kernels-mixer==0.0.7
382
+ keyring==24.3.0
383
+ keyrings.google-artifactregistry-auth==1.1.2
384
+ kfp-pipeline-spec==0.2.2
385
+ kfp-server-api==2.0.5
386
+ kfp==2.5.0
387
+ kiwisolver==1.4.5
388
+ kmapper==2.0.1
389
+ kmodes==0.12.2
390
+ korean-lunar-calendar==0.3.1
391
+ kornia==0.7.2
392
+ kornia_rs==0.1.3
393
+ kt-legacy==1.0.5
394
+ kubernetes==26.1.0
395
+ langcodes==3.3.0
396
+ langid==1.1.6
397
+ lazy_loader==0.3
398
+ learntools==0.3.4
399
+ leven==1.0.4
400
+ libclang==16.0.6
401
+ libmambapy==1.5.0
402
+ libpysal==4.9.2
403
+ librosa==0.10.1
404
+ lightgbm==4.2.0
405
+ lightning-utilities==0.11.2
406
+ lime==0.2.0.1
407
+ line-profiler==4.1.2
408
+ linkify-it-py==2.0.3
409
+ llvmlite==0.41.1
410
+ llvmlite==0.42.0
411
+ lml==0.1.0
412
+ locket==1.0.0
413
+ loguru==0.7.2
414
+ lxml==5.2.1
415
+ lz4==4.3.3
416
+ mamba==1.5.0
417
+ mapclassify==2.6.1
418
+ markdown-it-py==3.0.0
419
+ marshmallow==3.21.1
420
+ matplotlib-inline==0.1.6
421
+ matplotlib-venn==0.11.10
422
+ matplotlib==3.7.5
423
+ matplotlib==3.8.4
424
+ mccabe==0.7.0
425
+ mdit-py-plugins==0.4.0
426
+ mdurl==0.1.2
427
+ memory-profiler==0.61.0
428
+ menuinst==2.0.1
429
+ mercantile==1.2.1
430
+ mgwr==2.2.1
431
+ missingno==0.5.2
432
+ mistune==0.8.4
433
+ mizani==0.11.1
434
+ ml-dtypes==0.2.0
435
+ mlcrate==0.2.0
436
+ mlens==0.2.3
437
+ mlxtend==0.23.1
438
+ mne==1.6.1
439
+ mnist==0.2.2
440
+ momepy==0.7.0
441
+ more-itertools==10.2.0
442
+ mpld3==0.5.10
443
+ mpmath==1.3.0
444
+ msgpack==1.0.7
445
+ multidict==6.0.4
446
+ multimethod==1.10
447
+ multipledispatch==1.0.0
448
+ multiprocess==0.70.16
449
+ munkres==1.1.4
450
+ murmurhash==1.0.10
451
+ mypy-extensions==1.0.0
452
+ namex==0.0.8
453
+ nb-conda-kernels==2.3.1
454
+ nb_conda==2.2.1
455
+ nbclassic==1.0.0
456
+ nbclient==0.5.13
457
+ nbconvert==6.4.5
458
+ nbdime==3.2.0
459
+ nbformat==5.9.2
460
+ ndindex==1.8
461
+ nest-asyncio==1.5.8
462
+ networkx==3.2.1
463
+ nibabel==5.2.1
464
+ nilearn==0.10.4
465
+ ninja==1.11.1.1
466
+ nltk==3.2.4
467
+ nose==1.3.7
468
+ notebook==6.5.4
469
+ notebook==6.5.6
470
+ notebook_executor==0.2
471
+ notebook_shim==0.2.3
472
+ numba==0.58.1
473
+ numba==0.59.1
474
+ numexpr==2.10.0
475
+ numpy==1.26.4
476
+ nvidia-ml-py==11.495.46
477
+ nvtx==0.2.10
478
+ oauth2client==4.1.3
479
+ oauthlib==3.2.2
480
+ objsize==0.6.1
481
+ odfpy==1.4.1
482
+ olefile==0.47
483
+ onnx==1.16.0
484
+ opencensus-context==0.1.3
485
+ opencensus==0.11.4
486
+ opencv-contrib-python==4.9.0.80
487
+ opencv-python-headless==4.9.0.80
488
+ opencv-python==4.9.0.80
489
+ openpyxl==3.1.2
490
+ openslide-python==1.3.1
491
+ opentelemetry-api==1.22.0
492
+ opentelemetry-exporter-otlp-proto-common==1.22.0
493
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
494
+ opentelemetry-exporter-otlp-proto-http==1.22.0
495
+ opentelemetry-exporter-otlp==1.22.0
496
+ opentelemetry-proto==1.22.0
497
+ opentelemetry-sdk==1.22.0
498
+ opentelemetry-semantic-conventions==0.43b0
499
+ opt-einsum==3.3.0
500
+ optax==0.2.2
501
+ optree==0.11.0
502
+ optuna==3.6.1
503
+ orbax-checkpoint==0.5.9
504
+ ordered-set==4.1.0
505
+ orjson==3.9.10
506
+ ortools==9.4.1874
507
+ osmnx==1.9.2
508
+ overrides==7.4.0
509
+ packaging==21.3
510
+ pandas-datareader==0.10.0
511
+ pandas-profiling==3.6.6
512
+ pandas-summary==0.2.0
513
+ pandas==2.1.4
514
+ pandas==2.2.2
515
+ pandasql==0.7.3
516
+ pandocfilters==1.5.0
517
+ panel==1.4.1
518
+ papermill==2.5.0
519
+ param==2.1.0
520
+ parso==0.8.3
521
+ partd==1.4.1
522
+ path.py==12.5.0
523
+ path==16.14.0
524
+ pathos==0.3.2
525
+ pathy==0.10.3
526
+ patsy==0.5.6
527
+ pdf2image==1.17.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.21
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.7.0
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.18.1
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.4
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.2.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.2
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rapidfuzz==3.9.1
640
+ rasterio==1.3.10
641
+ rasterstats==0.19.0
642
+ ray-cpp==2.9.0
643
+ ray==2.9.0
644
+ referencing==0.32.1
645
+ regex==2023.12.25
646
+ requests-oauthlib==1.3.1
647
+ requests-toolbelt==0.10.1
648
+ requests==2.31.0
649
+ retrying==1.3.3
650
+ retrying==1.3.4
651
+ rfc3339-validator==0.1.4
652
+ rfc3986-validator==0.1.1
653
+ rgf-python==3.12.0
654
+ rich-click==1.7.4
655
+ rich==13.7.0
656
+ rich==13.7.1
657
+ rmm==23.8.0
658
+ rope==1.13.0
659
+ rpds-py==0.16.2
660
+ rsa==4.9
661
+ ruamel-yaml-conda==0.15.100
662
+ ruamel.yaml.clib==0.2.7
663
+ ruamel.yaml==0.17.40
664
+ s2sphere==0.2.5
665
+ s3fs==2024.2.0
666
+ s3transfer==0.6.2
667
+ safetensors==0.4.3
668
+ scattertext==0.1.19
669
+ scikit-image==0.22.0
670
+ scikit-learn-intelex==2024.3.0
671
+ scikit-learn==1.2.2
672
+ scikit-multilearn==0.2.0
673
+ scikit-optimize==0.10.1
674
+ scikit-plot==0.3.7
675
+ scikit-surprise==1.1.3
676
+ scipy==1.11.4
677
+ scipy==1.13.0
678
+ seaborn==0.12.2
679
+ segment_anything==1.0
680
+ segregation==2.5
681
+ semver==3.0.2
682
+ sentencepiece==0.2.0
683
+ sentry-sdk==1.45.0
684
+ setproctitle==1.3.3
685
+ setuptools-git==1.2
686
+ setuptools-scm==8.0.4
687
+ setuptools==69.0.3
688
+ shap==0.44.1
689
+ shapely==2.0.4
690
+ shellingham==1.5.4
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.3
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ truststore==0.8.0
790
+ trx-python==0.2.9
791
+ tsfresh==0.20.2
792
+ typeguard==4.1.5
793
+ typer==0.9.0
794
+ typer==0.9.4
795
+ types-python-dateutil==2.8.19.20240106
796
+ typing-inspect==0.9.0
797
+ typing-utils==0.1.0
798
+ typing_extensions==4.9.0
799
+ tzdata==2023.4
800
+ uc-micro-py==1.0.3
801
+ ucx-py==0.33.0
802
+ ujson==5.9.0
803
+ umap-learn==0.5.6
804
+ unicodedata2==15.1.0
805
+ update-checker==0.18.0
806
+ uri-template==1.3.0
807
+ uritemplate==3.0.1
808
+ urllib3==1.26.18
809
+ urllib3==2.1.0
810
+ urwid==2.6.10
811
+ urwid_readline==0.14
812
+ uvicorn==0.25.0
813
+ uvloop==0.19.0
814
+ vaex-astro==0.9.3
815
+ vaex-core==4.17.1
816
+ vaex-hdf5==0.14.1
817
+ vaex-jupyter==0.8.2
818
+ vaex-ml==0.18.3
819
+ vaex-server==0.9.0
820
+ vaex-viz==0.5.4
821
+ vaex==4.17.0
822
+ vec_noise==1.1.4
823
+ vecstack==0.4.0
824
+ virtualenv==20.21.0
825
+ visions==0.7.5
826
+ vowpalwabbit==9.9.0
827
+ vtk==9.3.0
828
+ wandb==0.16.6
829
+ wasabi==1.1.2
830
+ watchfiles==0.21.0
831
+ wavio==0.0.8
832
+ wcwidth==0.2.13
833
+ weasel==0.3.4
834
+ webcolors==1.13
835
+ webencodings==0.5.1
836
+ websocket-client==1.7.0
837
+ websockets==12.0
838
+ wfdb==4.1.2
839
+ whatthepatch==1.0.5
840
+ wheel==0.42.0
841
+ widgetsnbextension==3.6.6
842
+ witwidget==1.8.1
843
+ woodwork==0.30.0
844
+ wordcloud==1.9.3
845
+ wordsegment==1.3.1
846
+ wrapt==1.14.1
847
+ xarray-einstats==0.7.0
848
+ xarray==2024.3.0
849
+ xgboost==2.0.3
850
+ xvfbwrapper==0.2.9
851
+ xxhash==3.4.1
852
+ xyzservices==2024.4.0
853
+ y-py==0.6.2
854
+ yapf==0.40.2
855
+ yarl==1.9.3
856
+ yarl==1.9.4
857
+ ydata-profiling==4.6.4
858
+ yellowbrick==1.5
859
+ ypy-websocket==0.8.4
860
+ zict==3.0.0
861
+ zipp==3.17.0
862
+ zstandard==0.22.0
wandb/run-20240526_141305-wfyhopk0/files/wandb-metadata.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-05-26T14:13:06.561396",
5
+ "startedAt": "2024-05-26T14:13:05.022165",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "142d81d78b72",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.164,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.164,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.164,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.164,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.164,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5597.954666137695
49
+ }
50
+ },
51
+ "gpu": "Tesla P100-PCIE-16GB",
52
+ "gpu_count": 1,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla P100-PCIE-16GB",
56
+ "memory_total": 17179869184
57
+ }
58
+ ],
59
+ "memory": {
60
+ "total": 31.357559204101562
61
+ }
62
+ }
wandb/run-20240526_141305-wfyhopk0/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 4343.6755, "train/grad_norm": NaN, "train/learning_rate": 3.8e-06, "train/epoch": 5.68, "train/global_step": 500, "_timestamp": 1716733124.540537, "_runtime": 339.51146721839905, "_step": 1, "eval/loss": NaN, "eval/wer": 1.0, "eval/runtime": 10.0932, "eval/samples_per_second": 19.815, "eval/steps_per_second": 2.477}
wandb/run-20240526_141305-wfyhopk0/logs/debug-internal.log ADDED
@@ -0,0 +1,233 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-05-26 14:13:05,028 INFO StreamThr :225 [internal.py:wandb_internal():86] W&B internal server running at pid: 225, started at: 2024-05-26 14:13:05.028130
2
+ 2024-05-26 14:13:05,030 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status
3
+ 2024-05-26 14:13:06,044 INFO WriterThread:225 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240526_141305-wfyhopk0/run-wfyhopk0.wandb
4
+ 2024-05-26 14:13:06,044 DEBUG SenderThread:225 [sender.py:send():379] send: header
5
+ 2024-05-26 14:13:06,047 DEBUG SenderThread:225 [sender.py:send():379] send: run
6
+ 2024-05-26 14:13:06,468 INFO SenderThread:225 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240526_141305-wfyhopk0/files
7
+ 2024-05-26 14:13:06,469 INFO SenderThread:225 [sender.py:_start_run_threads():1124] run started: wfyhopk0 with start time 1716732785.02907
8
+ 2024-05-26 14:13:06,477 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-05-26 14:13:06,477 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: check_version
10
+ 2024-05-26 14:13:06,534 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-05-26 14:13:06,545 DEBUG HandlerThread:225 [system_info.py:__init__():26] System info init
12
+ 2024-05-26 14:13:06,545 DEBUG HandlerThread:225 [system_info.py:__init__():41] System info init done
13
+ 2024-05-26 14:13:06,545 INFO HandlerThread:225 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-05-26 14:13:06,546 INFO SystemMonitor:225 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-05-26 14:13:06,546 INFO HandlerThread:225 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-05-26 14:13:06,546 INFO SystemMonitor:225 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-05-26 14:13:06,547 INFO SystemMonitor:225 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-05-26 14:13:06,549 INFO SystemMonitor:225 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-05-26 14:13:06,550 INFO SystemMonitor:225 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-05-26 14:13:06,550 INFO SystemMonitor:225 [interfaces.py:start():190] Started network monitoring
21
+ 2024-05-26 14:13:06,561 DEBUG HandlerThread:225 [system_info.py:probe():150] Probing system
22
+ 2024-05-26 14:13:06,563 DEBUG HandlerThread:225 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-05-26 14:13:06,563 DEBUG HandlerThread:225 [system_info.py:probe():198] Probing system done
24
+ 2024-05-26 14:13:06,563 DEBUG HandlerThread:225 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-05-26T14:13:06.561396', 'startedAt': '2024-05-26T14:13:05.022165', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '142d81d78b72', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.164, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.164, 'min': 0.0, 'max': 0.0}, {'current': 2000.164, 'min': 0.0, 'max': 0.0}, {'current': 2000.164, 'min': 0.0, 'max': 0.0}, {'current': 2000.164, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5597.954666137695}}, 'gpu': 'Tesla P100-PCIE-16GB', 'gpu_count': 1, 'gpu_devices': [{'name': 'Tesla P100-PCIE-16GB', 'memory_total': 17179869184}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-05-26 14:13:06,563 INFO HandlerThread:225 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-05-26 14:13:06,563 INFO HandlerThread:225 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-05-26 14:13:06,563 DEBUG HandlerThread:225 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-05-26 14:13:07,470 INFO Thread-12 :225 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240526_141305-wfyhopk0/files/conda-environment.yaml
29
+ 2024-05-26 14:13:21,580 ERROR HandlerThread:225 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-05-26 14:13:21,583 DEBUG HandlerThread:225 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-05-26 14:13:21,583 INFO HandlerThread:225 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-05-26 14:13:21,591 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-05-26 14:13:21,592 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-05-26 14:13:21,592 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-05-26 14:13:21,592 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-05-26 14:13:21,592 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-05-26 14:13:21,592 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-05-26 14:13:21,592 DEBUG SenderThread:225 [sender.py:send():379] send: files
49
+ 2024-05-26 14:13:21,593 INFO SenderThread:225 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-05-26 14:13:22,239 INFO wandb-upload_0:225 [upload_job.py:push():131] Uploaded file /tmp/tmpfxjhfb1zwandb/k20mz4b6-wandb-metadata.json
51
+ 2024-05-26 14:13:22,474 INFO Thread-12 :225 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240526_141305-wfyhopk0/files/wandb-metadata.json
52
+ 2024-05-26 14:13:22,557 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-05-26 14:13:22,558 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-05-26 14:13:22,560 DEBUG SenderThread:225 [sender.py:send():379] send: telemetry
55
+ 2024-05-26 14:13:22,573 DEBUG SenderThread:225 [sender.py:send():379] send: config
56
+ 2024-05-26 14:13:22,575 DEBUG SenderThread:225 [sender.py:send():379] send: metric
57
+ 2024-05-26 14:13:22,575 DEBUG SenderThread:225 [sender.py:send():379] send: telemetry
58
+ 2024-05-26 14:13:22,575 DEBUG SenderThread:225 [sender.py:send():379] send: metric
59
+ 2024-05-26 14:13:22,575 WARNING SenderThread:225 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
60
+ 2024-05-26 14:13:22,575 DEBUG SenderThread:225 [sender.py:send():379] send: telemetry
61
+ 2024-05-26 14:13:22,576 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
62
+ 2024-05-26 14:13:22,577 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
63
+ 2024-05-26 14:13:22,577 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
64
+ 2024-05-26 14:13:23,474 INFO Thread-12 :225 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240526_141305-wfyhopk0/files/requirements.txt
65
+ 2024-05-26 14:13:23,475 INFO Thread-12 :225 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240526_141305-wfyhopk0/files/output.log
66
+ 2024-05-26 14:13:25,475 INFO Thread-12 :225 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240526_141305-wfyhopk0/files/output.log
67
+ 2024-05-26 14:13:25,533 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-05-26 14:13:30,534 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
69
+ 2024-05-26 14:13:35,540 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
70
+ 2024-05-26 14:13:36,479 INFO Thread-12 :225 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240526_141305-wfyhopk0/files/config.yaml
71
+ 2024-05-26 14:13:37,729 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
72
+ 2024-05-26 14:13:37,730 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
73
+ 2024-05-26 14:13:37,732 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
74
+ 2024-05-26 14:13:40,943 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
75
+ 2024-05-26 14:13:45,944 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
76
+ 2024-05-26 14:13:50,945 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
77
+ 2024-05-26 14:13:52,677 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
78
+ 2024-05-26 14:13:52,677 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
79
+ 2024-05-26 14:13:52,717 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
80
+ 2024-05-26 14:13:56,897 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-05-26 14:14:01,898 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
82
+ 2024-05-26 14:14:06,550 DEBUG SystemMonitor:225 [system_monitor.py:_start():172] Starting system metrics aggregation loop
83
+ 2024-05-26 14:14:06,552 DEBUG SenderThread:225 [sender.py:send():379] send: stats
84
+ 2024-05-26 14:14:07,553 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
85
+ 2024-05-26 14:14:07,975 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
86
+ 2024-05-26 14:14:07,976 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
87
+ 2024-05-26 14:14:07,977 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
88
+ 2024-05-26 14:14:13,194 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
89
+ 2024-05-26 14:14:18,196 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
90
+ 2024-05-26 14:14:22,965 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
91
+ 2024-05-26 14:14:22,974 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
92
+ 2024-05-26 14:14:22,975 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
93
+ 2024-05-26 14:14:24,158 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
94
+ 2024-05-26 14:14:29,159 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
95
+ 2024-05-26 14:14:34,160 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
96
+ 2024-05-26 14:14:36,553 DEBUG SenderThread:225 [sender.py:send():379] send: stats
97
+ 2024-05-26 14:14:37,966 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
98
+ 2024-05-26 14:14:37,975 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
99
+ 2024-05-26 14:14:37,975 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
100
+ 2024-05-26 14:14:39,207 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
101
+ 2024-05-26 14:14:44,208 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
102
+ 2024-05-26 14:14:49,210 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
103
+ 2024-05-26 14:14:53,045 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
104
+ 2024-05-26 14:14:53,066 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
105
+ 2024-05-26 14:14:53,067 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
106
+ 2024-05-26 14:14:54,265 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
107
+ 2024-05-26 14:14:59,266 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
108
+ 2024-05-26 14:15:04,267 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
109
+ 2024-05-26 14:15:06,554 DEBUG SenderThread:225 [sender.py:send():379] send: stats
110
+ 2024-05-26 14:15:08,034 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
111
+ 2024-05-26 14:15:08,034 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
112
+ 2024-05-26 14:15:08,035 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
113
+ 2024-05-26 14:15:10,176 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
114
+ 2024-05-26 14:15:15,177 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
115
+ 2024-05-26 14:15:20,179 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
116
+ 2024-05-26 14:15:23,224 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
117
+ 2024-05-26 14:15:23,225 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
118
+ 2024-05-26 14:15:23,228 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
119
+ 2024-05-26 14:15:25,427 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
120
+ 2024-05-26 14:15:30,428 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
121
+ 2024-05-26 14:15:35,428 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
122
+ 2024-05-26 14:15:36,555 DEBUG SenderThread:225 [sender.py:send():379] send: stats
123
+ 2024-05-26 14:15:38,194 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
124
+ 2024-05-26 14:15:38,195 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
125
+ 2024-05-26 14:15:38,195 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
126
+ 2024-05-26 14:15:41,356 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
127
+ 2024-05-26 14:15:46,356 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
128
+ 2024-05-26 14:15:51,357 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
129
+ 2024-05-26 14:15:53,194 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
130
+ 2024-05-26 14:15:53,195 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
131
+ 2024-05-26 14:15:53,195 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
132
+ 2024-05-26 14:15:56,402 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
133
+ 2024-05-26 14:16:01,403 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
134
+ 2024-05-26 14:16:06,404 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
135
+ 2024-05-26 14:16:06,555 DEBUG SenderThread:225 [sender.py:send():379] send: stats
136
+ 2024-05-26 14:16:08,194 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
137
+ 2024-05-26 14:16:08,195 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
138
+ 2024-05-26 14:16:08,195 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
139
+ 2024-05-26 14:16:11,441 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
140
+ 2024-05-26 14:16:16,442 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
141
+ 2024-05-26 14:16:21,443 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
142
+ 2024-05-26 14:16:23,331 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
143
+ 2024-05-26 14:16:23,332 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
144
+ 2024-05-26 14:16:23,332 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
145
+ 2024-05-26 14:16:26,538 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
146
+ 2024-05-26 14:16:31,539 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
147
+ 2024-05-26 14:16:36,540 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
148
+ 2024-05-26 14:16:36,556 DEBUG SenderThread:225 [sender.py:send():379] send: stats
149
+ 2024-05-26 14:16:38,437 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
150
+ 2024-05-26 14:16:38,438 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
151
+ 2024-05-26 14:16:38,439 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
152
+ 2024-05-26 14:16:41,579 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
153
+ 2024-05-26 14:16:46,580 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
154
+ 2024-05-26 14:16:51,581 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
155
+ 2024-05-26 14:16:53,400 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
156
+ 2024-05-26 14:16:53,401 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
157
+ 2024-05-26 14:16:53,401 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
158
+ 2024-05-26 14:16:56,646 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
159
+ 2024-05-26 14:17:01,647 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
160
+ 2024-05-26 14:17:06,557 DEBUG SenderThread:225 [sender.py:send():379] send: stats
161
+ 2024-05-26 14:17:07,558 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
162
+ 2024-05-26 14:17:08,401 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
163
+ 2024-05-26 14:17:08,401 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
164
+ 2024-05-26 14:17:08,401 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
165
+ 2024-05-26 14:17:12,573 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
166
+ 2024-05-26 14:17:17,574 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
167
+ 2024-05-26 14:17:22,575 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
168
+ 2024-05-26 14:17:23,457 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
169
+ 2024-05-26 14:17:23,457 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
170
+ 2024-05-26 14:17:23,461 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
171
+ 2024-05-26 14:17:27,603 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
172
+ 2024-05-26 14:17:32,604 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
173
+ 2024-05-26 14:17:36,558 DEBUG SenderThread:225 [sender.py:send():379] send: stats
174
+ 2024-05-26 14:17:38,511 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
175
+ 2024-05-26 14:17:38,512 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
176
+ 2024-05-26 14:17:38,512 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
177
+ 2024-05-26 14:17:38,513 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
178
+ 2024-05-26 14:17:43,750 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
179
+ 2024-05-26 14:17:48,751 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
180
+ 2024-05-26 14:17:53,602 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
181
+ 2024-05-26 14:17:53,602 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
182
+ 2024-05-26 14:17:53,643 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
183
+ 2024-05-26 14:17:53,851 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
184
+ 2024-05-26 14:17:58,853 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
185
+ 2024-05-26 14:18:03,854 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
186
+ 2024-05-26 14:18:06,559 DEBUG SenderThread:225 [sender.py:send():379] send: stats
187
+ 2024-05-26 14:18:08,729 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
188
+ 2024-05-26 14:18:08,730 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
189
+ 2024-05-26 14:18:08,769 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
190
+ 2024-05-26 14:18:08,886 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
191
+ 2024-05-26 14:18:13,886 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
192
+ 2024-05-26 14:18:18,887 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
193
+ 2024-05-26 14:18:23,707 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
194
+ 2024-05-26 14:18:23,708 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
195
+ 2024-05-26 14:18:23,748 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
196
+ 2024-05-26 14:18:23,935 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
197
+ 2024-05-26 14:18:28,935 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
198
+ 2024-05-26 14:18:33,937 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
199
+ 2024-05-26 14:18:34,444 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: partial_history
200
+ 2024-05-26 14:18:34,446 DEBUG SenderThread:225 [sender.py:send():379] send: metric
201
+ 2024-05-26 14:18:34,446 DEBUG SenderThread:225 [sender.py:send():379] send: metric
202
+ 2024-05-26 14:18:34,447 DEBUG SenderThread:225 [sender.py:send():379] send: metric
203
+ 2024-05-26 14:18:34,447 DEBUG SenderThread:225 [sender.py:send():379] send: metric
204
+ 2024-05-26 14:18:34,447 DEBUG SenderThread:225 [sender.py:send():379] send: history
205
+ 2024-05-26 14:18:34,447 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: summary_record
206
+ 2024-05-26 14:18:34,447 INFO SenderThread:225 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
207
+ 2024-05-26 14:18:34,595 INFO Thread-12 :225 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240526_141305-wfyhopk0/files/wandb-summary.json
208
+ 2024-05-26 14:18:36,560 DEBUG SenderThread:225 [sender.py:send():379] send: stats
209
+ 2024-05-26 14:18:38,725 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
210
+ 2024-05-26 14:18:38,726 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
211
+ 2024-05-26 14:18:38,728 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
212
+ 2024-05-26 14:18:38,980 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
213
+ 2024-05-26 14:18:43,989 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
214
+ 2024-05-26 14:18:44,541 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: partial_history
215
+ 2024-05-26 14:18:44,543 DEBUG SenderThread:225 [sender.py:send():379] send: metric
216
+ 2024-05-26 14:18:44,543 DEBUG SenderThread:225 [sender.py:send():379] send: metric
217
+ 2024-05-26 14:18:44,544 DEBUG SenderThread:225 [sender.py:send():379] send: metric
218
+ 2024-05-26 14:18:44,544 DEBUG SenderThread:225 [sender.py:send():379] send: metric
219
+ 2024-05-26 14:18:44,544 DEBUG SenderThread:225 [sender.py:send():379] send: metric
220
+ 2024-05-26 14:18:44,544 DEBUG SenderThread:225 [sender.py:send():379] send: history
221
+ 2024-05-26 14:18:44,544 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: summary_record
222
+ 2024-05-26 14:18:44,545 INFO SenderThread:225 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
223
+ 2024-05-26 14:18:44,599 INFO Thread-12 :225 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240526_141305-wfyhopk0/files/config.yaml
224
+ 2024-05-26 14:18:44,599 INFO Thread-12 :225 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240526_141305-wfyhopk0/files/wandb-summary.json
225
+ 2024-05-26 14:18:47,600 INFO Thread-12 :225 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240526_141305-wfyhopk0/files/output.log
226
+ 2024-05-26 14:18:49,772 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
227
+ 2024-05-26 14:18:53,717 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: stop_status
228
+ 2024-05-26 14:18:53,717 DEBUG SenderThread:225 [sender.py:send_request():406] send_request: stop_status
229
+ 2024-05-26 14:18:53,728 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: internal_messages
230
+ 2024-05-26 14:18:54,981 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
231
+ 2024-05-26 14:18:59,983 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
232
+ 2024-05-26 14:19:04,983 DEBUG HandlerThread:225 [handler.py:handle_request():146] handle_request: status_report
233
+ 2024-05-26 14:19:06,561 DEBUG SenderThread:225 [sender.py:send():379] send: stats
wandb/run-20240526_141305-wfyhopk0/logs/debug.log ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-05-26 14:13:05,023 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.6
2
+ 2024-05-26 14:13:05,023 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
7
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
8
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
10
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging user logs to /kaggle/working/wandb/run-20240526_141305-wfyhopk0/logs/debug.log
11
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_init.py:_log_setup():522] Logging internal logs to /kaggle/working/wandb/run-20240526_141305-wfyhopk0/logs/debug-internal.log
12
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_init.py:_jupyter_setup():467] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7bfadacb7c10>
13
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_init.py:init():561] calling init triggers
14
+ 2024-05-26 14:13:05,024 INFO MainThread:34 [wandb_init.py:init():568] wandb.init called with sweep_config: {}
15
+ config: {}
16
+ 2024-05-26 14:13:05,025 INFO MainThread:34 [wandb_init.py:init():611] starting backend
17
+ 2024-05-26 14:13:05,025 INFO MainThread:34 [wandb_init.py:init():615] setting up manager
18
+ 2024-05-26 14:13:05,026 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-05-26 14:13:05,028 INFO MainThread:34 [wandb_init.py:init():623] backend started and connected
20
+ 2024-05-26 14:13:05,040 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1299] probe notebook
21
+ 2024-05-26 14:13:06,043 INFO MainThread:34 [wandb_init.py:init():715] updated telemetry
22
+ 2024-05-26 14:13:06,046 INFO MainThread:34 [wandb_init.py:init():748] communicating run to backend with 90.0 second timeout
23
+ 2024-05-26 14:13:06,476 INFO MainThread:34 [wandb_run.py:_on_init():2357] communicating current version
24
+ 2024-05-26 14:13:06,526 INFO MainThread:34 [wandb_run.py:_on_init():2366] got version response upgrade_message: "wandb version 0.17.0 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
+
26
+ 2024-05-26 14:13:06,527 INFO MainThread:34 [wandb_init.py:init():799] starting run threads in backend
27
+ 2024-05-26 14:13:22,559 INFO MainThread:34 [wandb_run.py:_console_start():2335] atexit reg
28
+ 2024-05-26 14:13:22,559 INFO MainThread:34 [wandb_run.py:_redirect():2190] redirect: wrap_raw
29
+ 2024-05-26 14:13:22,559 INFO MainThread:34 [wandb_run.py:_redirect():2255] Wrapping output streams.
30
+ 2024-05-26 14:13:22,559 INFO MainThread:34 [wandb_run.py:_redirect():2280] Redirects installed.
31
+ 2024-05-26 14:13:22,560 INFO MainThread:34 [wandb_init.py:init():842] run started, returning control to user process
32
+ 2024-05-26 14:13:22,567 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 100000000000000, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May26_14-12-33_142d81d78b72', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240526_141305-wfyhopk0/run-wfyhopk0.wandb ADDED
Binary file (15 kB). View file