download
history
blame
contribute
delete
58.9 MB
No virus
Detected Pickle imports (166)
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1500.Linear",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1485.BertAttention",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1465.BertIntermediate",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1442.BertIntermediate",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1463.LinearPackedParams",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1456.BertSelfAttention",
- "collections.OrderedDict",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1508.BertAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1469.Dropout",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1498.Linear",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_1404.Embedding",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1432.Dropout",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1431.Linear",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1487.Linear",
- "__torch__.sentence_transformers.models.Normalize.___torch_mangle_1547.Normalize",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1453.LinearPackedParams",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1457.LinearPackedParams",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1414.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1440.LinearPackedParams",
- "__torch__.sentence_transformers.SentenceTransformer.___torch_mangle_1548.SentenceTransformer",
- "__torch__.sentence_transformers.models.Pooling.___torch_mangle_1546.Pooling",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1436.LayerNorm",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1473.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1445.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1492.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1459.LayerNorm",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1517.BertLayer",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1495.LinearPackedParams",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1420.LinearPackedParams",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1494.BertLayer",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1493.BertOutput",
- "torch.LongStorage",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1462.BertAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1460.Dropout",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1539.BertOutput",
- "torch._utils._rebuild_qtensor",
- "__torch__.transformers.models.bert.modeling_bert.BertEmbeddings",
- "__torch__.transformers.models.bert.modeling_bert.BertAttention",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1502.BertSelfAttention",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1484.BertSelfOutput",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1507.BertSelfOutput",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1542.LinearPackedParams",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1518.LinearPackedParams",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1543.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1409.LinearPackedParams",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1536.Linear",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1471.BertLayer",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1467.Linear",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1441.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1483.Dropout",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1443.LinearPackedParams",
- "__torch__.transformers.models.bert.modeling_bert.BertIntermediate",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1522.LinearPackedParams",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1447.BertOutput",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1421.Linear",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1452.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1509.LinearPackedParams",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1424.LayerNorm",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1423.Linear",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1513.Linear",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1470.BertOutput",
- "torch.per_tensor_affine",
- "__torch__.sentence_transformers.models.Transformer.___torch_mangle_1545.Transformer",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1450.Linear",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1439.BertAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1538.Dropout",
- "__torch__.transformers.models.bert.modeling_bert.BertLayer",
- "torch.FloatStorage",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1461.BertSelfOutput",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1464.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1491.LayerNorm",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1519.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1526.LinearPackedParams",
- "__torch__.transformers.models.bert.modeling_bert.BertSelfOutput",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1534.BertIntermediate",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1533.Linear",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1477.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1489.LinearPackedParams",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1422.LinearPackedParams",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1523.Linear",
- "__torch__.transformers.models.bert.modeling_bert.BertEncoder",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1521.Linear",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1448.BertLayer",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1510.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1426.LinearPackedParams",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1490.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1455.Dropout",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1527.Linear",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.transformers.models.bert.modeling_bert.BertOutput",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1413.LinearPackedParams",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1475.Linear",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1410.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1474.LinearPackedParams",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1429.Linear",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1438.BertSelfOutput",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1478.Dropout",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1449.LinearPackedParams",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1468.LayerNorm",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1444.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1472.LinearPackedParams",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1416.LinearPackedParams",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1425.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1482.LayerNorm",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1434.LinearPackedParams",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_1406.Embedding",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1480.LinearPackedParams",
- "__torch__.transformers.models.bert.modeling_bert.BertPooler",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1529.Dropout",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1503.LinearPackedParams",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1479.BertSelfAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1418.LayerNorm",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1488.BertIntermediate",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1499.LinearPackedParams",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1505.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1408.Dropout",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1516.BertOutput",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1446.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1514.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1506.Dropout",
- "__torch__.transformers.models.bert.modeling_bert.BertModel",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1411.LinearPackedParams",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1476.LinearPackedParams",
- "__torch__.torch.nn.modules.container.___torch_mangle_1541.ModuleList",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1504.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1532.LinearPackedParams",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1512.LinearPackedParams",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1486.LinearPackedParams",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1458.Linear",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1530.BertSelfOutput",
- "torch.QInt8Storage",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1412.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1520.LinearPackedParams",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1524.Dropout",
- "__torch__.transformers.models.bert.modeling_bert.BertSelfAttention",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1417.Linear",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1433.BertSelfAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1407.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1419.Dropout",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1496.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1535.LinearPackedParams",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1537.LayerNorm",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1451.LinearPackedParams",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1531.BertAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1415.Dropout",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1497.LinearPackedParams",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1511.BertIntermediate",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1525.BertSelfAttention",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1430.LinearPackedParams",
- "__torch__.torch.nn.modules.activation.___torch_mangle_1544.Tanh",
- "__torch__.transformers.models.bert.modeling_bert.___torch_mangle_1540.BertLayer",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1437.Dropout",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1481.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1428.LinearPackedParams",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1501.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1528.LayerNorm",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1427.Linear",
- "__torch__.torch.nn.quantized.modules.linear.___torch_mangle_1466.LinearPackedParams",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1435.Linear",
- "__torch__.torch.nn.quantized.dynamic.modules.linear.___torch_mangle_1454.Linear",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_1405.Embedding",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1515.Dropout",
- "torch.LongStorage",
- "collections.OrderedDict",
- "torch._utils._rebuild_tensor_v2",
- "torch.DoubleStorage"
Git LFS Details
- SHA256: dfcbf5dae96489123e8e9b2c6b2f70e69ed8216cdc4b9201263a9ca3a91176fb
- Pointer size: 133 Bytes
- Size of remote file: 58.9 MB
Git Large File Storage (LFS) replaces large files with text pointers inside Git, while storing the file contents on a remote server. More info.